var/home/core/zuul-output/0000755000175000017500000000000015111126665014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111146461015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005334003315111146451017675 0ustar rootrootNov 24 19:14:20 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 19:14:20 crc restorecon[4741]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:20 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 19:14:21 crc restorecon[4741]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 19:14:21 crc kubenswrapper[5035]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 19:14:21 crc kubenswrapper[5035]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 19:14:21 crc kubenswrapper[5035]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 19:14:21 crc kubenswrapper[5035]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 19:14:21 crc kubenswrapper[5035]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 19:14:21 crc kubenswrapper[5035]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.943253 5035 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950265 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950372 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950383 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950395 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950407 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950453 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950481 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950492 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950502 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950517 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950546 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950557 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950569 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950579 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950589 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950599 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950610 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950620 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950634 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950651 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950663 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950675 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950696 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950708 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950719 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950730 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950741 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950806 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950832 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950845 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950858 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950869 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950880 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950893 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950904 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950912 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950921 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950929 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950938 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950955 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950969 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950979 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.950990 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951000 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951008 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951017 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951026 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951034 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951045 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951053 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951072 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951082 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951090 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951098 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951106 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951114 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951123 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951131 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951139 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951147 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951157 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951164 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951178 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951186 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951194 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951202 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951210 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951218 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951226 5035 feature_gate.go:330] unrecognized feature gate: Example Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951234 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.951241 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951569 5035 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951889 5035 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951908 5035 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951921 5035 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951933 5035 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951943 5035 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951956 5035 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951968 5035 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951977 5035 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951987 5035 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.951997 5035 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952006 5035 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952016 5035 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952026 5035 flags.go:64] FLAG: --cgroup-root="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952035 5035 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952045 5035 flags.go:64] FLAG: --client-ca-file="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952054 5035 flags.go:64] FLAG: --cloud-config="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952064 5035 flags.go:64] FLAG: --cloud-provider="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952073 5035 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952086 5035 flags.go:64] FLAG: --cluster-domain="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952094 5035 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952104 5035 flags.go:64] FLAG: --config-dir="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952113 5035 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952124 5035 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952135 5035 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952144 5035 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952154 5035 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952163 5035 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952173 5035 flags.go:64] FLAG: --contention-profiling="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952182 5035 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952191 5035 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952201 5035 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952209 5035 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952222 5035 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952231 5035 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952240 5035 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952249 5035 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952258 5035 flags.go:64] FLAG: --enable-server="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952268 5035 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952281 5035 flags.go:64] FLAG: --event-burst="100" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952358 5035 flags.go:64] FLAG: --event-qps="50" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952370 5035 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952379 5035 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952388 5035 flags.go:64] FLAG: --eviction-hard="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952410 5035 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952419 5035 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952428 5035 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952437 5035 flags.go:64] FLAG: --eviction-soft="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952446 5035 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952456 5035 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952465 5035 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952474 5035 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952483 5035 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952492 5035 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952501 5035 flags.go:64] FLAG: --feature-gates="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952513 5035 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952523 5035 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952532 5035 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952541 5035 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952551 5035 flags.go:64] FLAG: --healthz-port="10248" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952560 5035 flags.go:64] FLAG: --help="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952570 5035 flags.go:64] FLAG: --hostname-override="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952579 5035 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952588 5035 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952599 5035 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952608 5035 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952616 5035 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952625 5035 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952634 5035 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952643 5035 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952652 5035 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952662 5035 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952672 5035 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952681 5035 flags.go:64] FLAG: --kube-reserved="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952690 5035 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952700 5035 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952709 5035 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952718 5035 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952727 5035 flags.go:64] FLAG: --lock-file="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952736 5035 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952746 5035 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952758 5035 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952773 5035 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952784 5035 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952794 5035 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952804 5035 flags.go:64] FLAG: --logging-format="text" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952813 5035 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952823 5035 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952831 5035 flags.go:64] FLAG: --manifest-url="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952840 5035 flags.go:64] FLAG: --manifest-url-header="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952854 5035 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952864 5035 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952875 5035 flags.go:64] FLAG: --max-pods="110" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952884 5035 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952893 5035 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952902 5035 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952912 5035 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952921 5035 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952930 5035 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952940 5035 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952963 5035 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952973 5035 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952982 5035 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952991 5035 flags.go:64] FLAG: --pod-cidr="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.952999 5035 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953013 5035 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953022 5035 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953031 5035 flags.go:64] FLAG: --pods-per-core="0" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953040 5035 flags.go:64] FLAG: --port="10250" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953049 5035 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953058 5035 flags.go:64] FLAG: --provider-id="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953068 5035 flags.go:64] FLAG: --qos-reserved="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953077 5035 flags.go:64] FLAG: --read-only-port="10255" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953087 5035 flags.go:64] FLAG: --register-node="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953096 5035 flags.go:64] FLAG: --register-schedulable="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953105 5035 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953121 5035 flags.go:64] FLAG: --registry-burst="10" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953130 5035 flags.go:64] FLAG: --registry-qps="5" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953141 5035 flags.go:64] FLAG: --reserved-cpus="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953151 5035 flags.go:64] FLAG: --reserved-memory="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953162 5035 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953172 5035 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953181 5035 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953190 5035 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953199 5035 flags.go:64] FLAG: --runonce="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953208 5035 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953217 5035 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953227 5035 flags.go:64] FLAG: --seccomp-default="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953236 5035 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953244 5035 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953254 5035 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953263 5035 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953272 5035 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953282 5035 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953316 5035 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953325 5035 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953334 5035 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953344 5035 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953353 5035 flags.go:64] FLAG: --system-cgroups="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953362 5035 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953377 5035 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953386 5035 flags.go:64] FLAG: --tls-cert-file="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953395 5035 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953406 5035 flags.go:64] FLAG: --tls-min-version="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953415 5035 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953424 5035 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953433 5035 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953442 5035 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953452 5035 flags.go:64] FLAG: --v="2" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953464 5035 flags.go:64] FLAG: --version="false" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953477 5035 flags.go:64] FLAG: --vmodule="" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953488 5035 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.953498 5035 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953723 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953733 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953744 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953754 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953762 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953771 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953778 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953786 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953794 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953802 5035 feature_gate.go:330] unrecognized feature gate: Example Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953809 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953817 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953825 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953832 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953841 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953850 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953858 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953867 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953874 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953882 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953890 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953897 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953905 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953912 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953920 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953928 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953937 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953944 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953952 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953960 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953968 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953978 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953986 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.953993 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954001 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954009 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954017 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954025 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954034 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954041 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954049 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954057 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954065 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954072 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954082 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954091 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954100 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954108 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954116 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954123 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954131 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954139 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954147 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954155 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954162 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954170 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954178 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954189 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954199 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954207 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954216 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954225 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954236 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954245 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954255 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954263 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954273 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954283 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954326 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954347 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.954360 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.955273 5035 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.968642 5035 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.968688 5035 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968802 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968815 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968824 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968832 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968841 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968850 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968859 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968867 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968875 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968883 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968890 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968898 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968906 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968916 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968924 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968931 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968939 5035 feature_gate.go:330] unrecognized feature gate: Example Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968947 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968955 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968962 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968970 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968981 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968990 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.968999 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969007 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969015 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969022 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969030 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969038 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969047 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969056 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969065 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969073 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969084 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969094 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969102 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969111 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969120 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969128 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969137 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969145 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969153 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969160 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969168 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969176 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969184 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969192 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969199 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969207 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969215 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969223 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969230 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969240 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969249 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969257 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969265 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969273 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969281 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969323 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969335 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969345 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969354 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969361 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969369 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969377 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969387 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969398 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969407 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969415 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969423 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969432 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.969445 5035 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969702 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969714 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969723 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969732 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969740 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969750 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969760 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969768 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969776 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969783 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969791 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969799 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969807 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969815 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969823 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969831 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969839 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969847 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969857 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969865 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969873 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969881 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969889 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969897 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969905 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969913 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969921 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969928 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969936 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969945 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969953 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969961 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969971 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969981 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.969992 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970002 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970012 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970021 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970030 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970038 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970046 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970054 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970065 5035 feature_gate.go:330] unrecognized feature gate: Example Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970073 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970081 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970090 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970097 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970105 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970113 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970121 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970129 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970139 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970148 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970157 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970164 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970173 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970180 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970188 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970199 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970208 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970217 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970226 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970233 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970241 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970249 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970258 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970266 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970274 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970282 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970320 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 19:14:21 crc kubenswrapper[5035]: W1124 19:14:21.970333 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.970349 5035 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.971799 5035 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.977371 5035 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.977507 5035 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.979495 5035 server.go:997] "Starting client certificate rotation" Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.979535 5035 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.980848 5035 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-11 03:47:35.997513422 +0000 UTC Nov 24 19:14:21 crc kubenswrapper[5035]: I1124 19:14:21.980968 5035 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1136h33m14.01655036s for next certificate rotation Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.009138 5035 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.011819 5035 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.030604 5035 log.go:25] "Validated CRI v1 runtime API" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.073187 5035 log.go:25] "Validated CRI v1 image API" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.075696 5035 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.081348 5035 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-19-09-45-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.081385 5035 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.104133 5035 manager.go:217] Machine: {Timestamp:2025-11-24 19:14:22.101550903 +0000 UTC m=+0.624057230 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:17ceffe9-e870-43f4-80f9-b6f15e4ae84e BootID:397c8539-193a-45fe-895a-942e12dca119 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:20:ff:e4 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:20:ff:e4 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ef:a4:cb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:d1:b5:ea Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d7:dc:dc Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a2:6b:6e Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:8b:cf:f2 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1e:b0:d0:dc:c9:24 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:e2:9d:c1:ea:c3:4f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.104585 5035 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.104796 5035 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.106238 5035 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.106569 5035 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.106614 5035 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.106915 5035 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.106933 5035 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.107423 5035 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.107467 5035 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.107713 5035 state_mem.go:36] "Initialized new in-memory state store" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.107851 5035 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.112603 5035 kubelet.go:418] "Attempting to sync node with API server" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.112637 5035 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.112675 5035 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.112696 5035 kubelet.go:324] "Adding apiserver pod source" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.112714 5035 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.116922 5035 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.117967 5035 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 19:14:22 crc kubenswrapper[5035]: W1124 19:14:22.119630 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:22 crc kubenswrapper[5035]: W1124 19:14:22.119743 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.119856 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.119784 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.120455 5035 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122359 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122425 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122447 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122463 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122491 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122508 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122521 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122543 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122560 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122574 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122593 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.122606 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.123688 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.124475 5035 server.go:1280] "Started kubelet" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.124671 5035 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.124916 5035 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.125167 5035 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 19:14:22 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.127429 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.127778 5035 server.go:460] "Adding debug handlers to kubelet server" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.128491 5035 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.128545 5035 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.128601 5035 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 07:47:49.222917974 +0000 UTC Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.128687 5035 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.128790 5035 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.128808 5035 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.128895 5035 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.133997 5035 factory.go:55] Registering systemd factory Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.134077 5035 factory.go:221] Registration of the systemd container factory successfully Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.134710 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="200ms" Nov 24 19:14:22 crc kubenswrapper[5035]: W1124 19:14:22.134950 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.135051 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.136174 5035 factory.go:153] Registering CRI-O factory Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.136308 5035 factory.go:221] Registration of the crio container factory successfully Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.136474 5035 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.136558 5035 factory.go:103] Registering Raw factory Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.136626 5035 manager.go:1196] Started watching for new ooms in manager Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.135648 5035 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.47:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b0742b4922a95 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 19:14:22.124395157 +0000 UTC m=+0.646901454,LastTimestamp:2025-11-24 19:14:22.124395157 +0000 UTC m=+0.646901454,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.137370 5035 manager.go:319] Starting recovery of all containers Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.145360 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.145544 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.145620 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.145683 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.145881 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.145956 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146027 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146115 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146185 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146255 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146368 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146429 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146485 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146550 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146611 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146693 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146754 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146823 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146893 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.146952 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147009 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147071 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147160 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147224 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147309 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147377 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147437 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.147495 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.148960 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.149035 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.149099 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.149158 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.149220 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.150611 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.150856 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151054 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151126 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151188 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151242 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151645 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151725 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151795 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151858 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151917 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.151990 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.152074 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.152142 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.152208 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.152269 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153587 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153726 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153757 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153790 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153897 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153950 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153974 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.153993 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154108 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154801 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154836 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154852 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154868 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154882 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154895 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154908 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154921 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154935 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154948 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154969 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154983 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.154997 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155010 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155025 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155039 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155052 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155065 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155078 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155099 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155112 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155127 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155140 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155153 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155167 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155181 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155424 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155440 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155454 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155470 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155483 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155498 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155512 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155526 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155539 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155555 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155568 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155582 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155596 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155609 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155624 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155640 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155654 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155670 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155684 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155697 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155717 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155732 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155747 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155762 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155776 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155793 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155809 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155826 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155840 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155854 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155868 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155881 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155895 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155908 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155930 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155948 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155965 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.155984 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156002 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156019 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156033 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156049 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156063 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156087 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156101 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156114 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156128 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156143 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156156 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156172 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156186 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156199 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156214 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156228 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156241 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156256 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156269 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156285 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156321 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156335 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156350 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156365 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156379 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156392 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156407 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156420 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156468 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156482 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156496 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156510 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156525 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156537 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156549 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156566 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156580 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156594 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156607 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156620 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156699 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156714 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156727 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156743 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156757 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156770 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156784 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156797 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156809 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156823 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156837 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156852 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156868 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156882 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156896 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156910 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.156927 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159226 5035 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159284 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159330 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159356 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159396 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159411 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159424 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159439 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159453 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159487 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159501 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159515 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159528 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159563 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159577 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159590 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159602 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159614 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159650 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159663 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159675 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159689 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159738 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159751 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159765 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159778 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159868 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159906 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159919 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159931 5035 reconstruct.go:97] "Volume reconstruction finished" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.159941 5035 reconciler.go:26] "Reconciler: start to sync state" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.164915 5035 manager.go:324] Recovery completed Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.176693 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.180416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.180489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.180501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.182234 5035 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.182279 5035 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.182324 5035 state_mem.go:36] "Initialized new in-memory state store" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.191331 5035 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.193621 5035 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.198791 5035 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.198841 5035 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.198968 5035 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 19:14:22 crc kubenswrapper[5035]: W1124 19:14:22.199510 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.199572 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.212001 5035 policy_none.go:49] "None policy: Start" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.213100 5035 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.213130 5035 state_mem.go:35] "Initializing new in-memory state store" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.229607 5035 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.256468 5035 manager.go:334] "Starting Device Plugin manager" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.256894 5035 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.256983 5035 server.go:79] "Starting device plugin registration server" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.257505 5035 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.257593 5035 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.257793 5035 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.257925 5035 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.257941 5035 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.264041 5035 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.299025 5035 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.299112 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.300240 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.300401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.300592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.300963 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.301196 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.301239 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.301988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.302013 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.302024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.303037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.303155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.303244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.303467 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.303587 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.303616 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304539 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304686 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304717 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.304719 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.305433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.305583 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.305452 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.305705 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.305728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.305662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.305965 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.306068 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.306139 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.307584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.307609 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.307622 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.307779 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.307870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.307948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.308171 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.308255 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.309131 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.309188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.309209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.335432 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="400ms" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.357803 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.359440 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.359520 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.359531 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.359575 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.360078 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361366 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361423 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361495 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361546 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361592 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361642 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361684 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361715 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361750 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361793 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361866 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361934 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.361973 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.362026 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463370 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463559 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463592 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463587 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463684 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463718 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463742 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463775 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463778 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463852 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463891 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463911 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463701 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463970 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.463972 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464015 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464100 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464069 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464185 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464224 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464252 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464284 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464361 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464368 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464446 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464509 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464690 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.464808 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.561041 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.563364 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.563414 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.563435 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.563477 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.565179 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.633604 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.658600 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.697239 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.697825 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.701247 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 19:14:22 crc kubenswrapper[5035]: W1124 19:14:22.707741 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9d855fc86bead9b0bf0837864e955783d1242c2be3698f843518fc386bbde955 WatchSource:0}: Error finding container 9d855fc86bead9b0bf0837864e955783d1242c2be3698f843518fc386bbde955: Status 404 returned error can't find the container with id 9d855fc86bead9b0bf0837864e955783d1242c2be3698f843518fc386bbde955 Nov 24 19:14:22 crc kubenswrapper[5035]: W1124 19:14:22.720591 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-27c8897a7c5271a8bc135e01b35fcdb34c44a7501d6b6136ac7f8f5e22ee8cb5 WatchSource:0}: Error finding container 27c8897a7c5271a8bc135e01b35fcdb34c44a7501d6b6136ac7f8f5e22ee8cb5: Status 404 returned error can't find the container with id 27c8897a7c5271a8bc135e01b35fcdb34c44a7501d6b6136ac7f8f5e22ee8cb5 Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.736926 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="800ms" Nov 24 19:14:22 crc kubenswrapper[5035]: W1124 19:14:22.738629 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f8288ecdb392129d5c8a1904e6bdbc7c2facc873ce0b94c1a50ff0d5fb02a839 WatchSource:0}: Error finding container f8288ecdb392129d5c8a1904e6bdbc7c2facc873ce0b94c1a50ff0d5fb02a839: Status 404 returned error can't find the container with id f8288ecdb392129d5c8a1904e6bdbc7c2facc873ce0b94c1a50ff0d5fb02a839 Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.965993 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.967244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.967310 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.967325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:22 crc kubenswrapper[5035]: I1124 19:14:22.967350 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 19:14:22 crc kubenswrapper[5035]: E1124 19:14:22.967765 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Nov 24 19:14:23 crc kubenswrapper[5035]: W1124 19:14:23.089097 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:23 crc kubenswrapper[5035]: E1124 19:14:23.089205 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.128472 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.129401 5035 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 19:36:49.82967676 +0000 UTC Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.129448 5035 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 96h22m26.700231921s for next certificate rotation Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.202885 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9d855fc86bead9b0bf0837864e955783d1242c2be3698f843518fc386bbde955"} Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.205145 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7dd1ef4c3f5f5c77697b3f870832cf360c605474d1a97939b3cd247680b9ba36"} Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.206160 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"72b3fc83ef4d170137608a77dcfa5b97cadb7f841ea5a781af6750c17d0ab2fc"} Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.207070 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f8288ecdb392129d5c8a1904e6bdbc7c2facc873ce0b94c1a50ff0d5fb02a839"} Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.207900 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"27c8897a7c5271a8bc135e01b35fcdb34c44a7501d6b6136ac7f8f5e22ee8cb5"} Nov 24 19:14:23 crc kubenswrapper[5035]: W1124 19:14:23.235945 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:23 crc kubenswrapper[5035]: E1124 19:14:23.236009 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:23 crc kubenswrapper[5035]: W1124 19:14:23.237969 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:23 crc kubenswrapper[5035]: E1124 19:14:23.238008 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:23 crc kubenswrapper[5035]: E1124 19:14:23.538467 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="1.6s" Nov 24 19:14:23 crc kubenswrapper[5035]: W1124 19:14:23.647842 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:23 crc kubenswrapper[5035]: E1124 19:14:23.647948 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.768384 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.769729 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.769784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.769800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:23 crc kubenswrapper[5035]: I1124 19:14:23.769838 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 19:14:23 crc kubenswrapper[5035]: E1124 19:14:23.770574 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.128632 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.212025 5035 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5" exitCode=0 Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.212104 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.212159 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.213070 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.213094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.213123 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.214912 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.214965 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.214986 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.215003 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.214924 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.216099 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.216121 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.216129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.217130 5035 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105" exitCode=0 Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.217208 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.217236 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.217862 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.217887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.217897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.218686 5035 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7" exitCode=0 Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.218749 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.218869 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.219895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.219942 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.219960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.220112 5035 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9ded3d4f659d102589c30d71475cd15ae8f4f93ced6dd7b6d83f41fd5846af38" exitCode=0 Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.220137 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9ded3d4f659d102589c30d71475cd15ae8f4f93ced6dd7b6d83f41fd5846af38"} Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.220188 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.220818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.220837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.220846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.221967 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.222956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.222972 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:24 crc kubenswrapper[5035]: I1124 19:14:24.222981 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:24 crc kubenswrapper[5035]: W1124 19:14:24.973259 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:24 crc kubenswrapper[5035]: E1124 19:14:24.973448 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.128851 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:25 crc kubenswrapper[5035]: E1124 19:14:25.141326 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="3.2s" Nov 24 19:14:25 crc kubenswrapper[5035]: W1124 19:14:25.222488 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Nov 24 19:14:25 crc kubenswrapper[5035]: E1124 19:14:25.222551 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.225700 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.225735 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.225747 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.225760 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.226627 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.226664 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.226675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.227593 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.227584 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.229775 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.229825 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.229888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.231595 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.231648 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.231667 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.231685 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.233783 5035 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ee52112ae3bbaf08412c464af26665e659b084d1c5de5682bb83c230355dec7a" exitCode=0 Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.233932 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.234155 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ee52112ae3bbaf08412c464af26665e659b084d1c5de5682bb83c230355dec7a"} Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.234272 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.235083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.235128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.235143 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.235088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.235211 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.235226 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.371084 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.372105 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.372133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.372142 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:25 crc kubenswrapper[5035]: I1124 19:14:25.372162 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 19:14:25 crc kubenswrapper[5035]: E1124 19:14:25.372679 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.239690 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be"} Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.239801 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.240561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.240577 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.240586 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.242989 5035 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5ce0c978c76ee9ee39d008de76021a67da21691cf6b69a8da39f2ff10a94999a" exitCode=0 Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.243074 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5ce0c978c76ee9ee39d008de76021a67da21691cf6b69a8da39f2ff10a94999a"} Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.243111 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.243161 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.243164 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.243224 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247125 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247215 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.247177 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.274169 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:26 crc kubenswrapper[5035]: I1124 19:14:26.691371 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.250494 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e71c2f28f22111584a5ef0ece180feb8f53e2ed2497986cce10a89b4507674fa"} Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.250598 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.250578 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.250576 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ed65b43fd95fa91dd434bfee31fe4783f633bd2ce68b7efe40f4d6e2dd58f5b5"} Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.250680 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d152850445416bfa2e6785604b84e25f3676ce0e2000225cc83b7bf35f59443c"} Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.250709 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8b5729f0a2dff4484c911d1f626a2af27a2194cb07cfa8583ade924a3a777045"} Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.251721 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.251746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.251779 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.251790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.251800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:27 crc kubenswrapper[5035]: I1124 19:14:27.251826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.259137 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a8c95781c31eb0f5eacae2ccc4e7fb5c1eeb9397b8243db3d0c6aafb63cd9d3e"} Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.259241 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.259259 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.260813 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.260866 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.260878 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.261007 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.261049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.261071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.573770 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.575224 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.575273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.575313 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.575342 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.588227 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.588517 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.590430 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.590462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:28 crc kubenswrapper[5035]: I1124 19:14:28.590476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:29 crc kubenswrapper[5035]: I1124 19:14:29.261682 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:29 crc kubenswrapper[5035]: I1124 19:14:29.262852 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:29 crc kubenswrapper[5035]: I1124 19:14:29.262883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:29 crc kubenswrapper[5035]: I1124 19:14:29.262893 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:29 crc kubenswrapper[5035]: I1124 19:14:29.424738 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.117119 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.117261 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.118144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.118169 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.118178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.263670 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.267347 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.267406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.267430 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.553949 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.554180 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.555772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.555988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:30 crc kubenswrapper[5035]: I1124 19:14:30.556161 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.020051 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.020355 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.021925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.021968 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.021983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.026779 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:32 crc kubenswrapper[5035]: E1124 19:14:32.264178 5035 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.266743 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.268038 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.268096 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.268112 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:32 crc kubenswrapper[5035]: I1124 19:14:32.525074 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:33 crc kubenswrapper[5035]: I1124 19:14:33.117548 5035 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 19:14:33 crc kubenswrapper[5035]: I1124 19:14:33.117668 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 19:14:33 crc kubenswrapper[5035]: I1124 19:14:33.269643 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:33 crc kubenswrapper[5035]: I1124 19:14:33.271219 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:33 crc kubenswrapper[5035]: I1124 19:14:33.271373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:33 crc kubenswrapper[5035]: I1124 19:14:33.271441 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:33 crc kubenswrapper[5035]: I1124 19:14:33.276151 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:34 crc kubenswrapper[5035]: I1124 19:14:34.272963 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:34 crc kubenswrapper[5035]: I1124 19:14:34.274015 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:34 crc kubenswrapper[5035]: I1124 19:14:34.274094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:34 crc kubenswrapper[5035]: I1124 19:14:34.274117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:35 crc kubenswrapper[5035]: W1124 19:14:35.858152 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 24 19:14:35 crc kubenswrapper[5035]: I1124 19:14:35.858259 5035 trace.go:236] Trace[1813648539]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 19:14:25.857) (total time: 10001ms): Nov 24 19:14:35 crc kubenswrapper[5035]: Trace[1813648539]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (19:14:35.858) Nov 24 19:14:35 crc kubenswrapper[5035]: Trace[1813648539]: [10.001187214s] [10.001187214s] END Nov 24 19:14:35 crc kubenswrapper[5035]: E1124 19:14:35.858286 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.129145 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.148657 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.148725 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.158280 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.158352 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.672873 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.673126 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.674335 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.674375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.674387 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:36 crc kubenswrapper[5035]: I1124 19:14:36.737598 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 19:14:37 crc kubenswrapper[5035]: I1124 19:14:37.676364 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:37 crc kubenswrapper[5035]: I1124 19:14:37.677284 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:37 crc kubenswrapper[5035]: I1124 19:14:37.677330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:37 crc kubenswrapper[5035]: I1124 19:14:37.677343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:37 crc kubenswrapper[5035]: I1124 19:14:37.688239 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 19:14:38 crc kubenswrapper[5035]: I1124 19:14:38.679555 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:38 crc kubenswrapper[5035]: I1124 19:14:38.681234 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:38 crc kubenswrapper[5035]: I1124 19:14:38.681349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:38 crc kubenswrapper[5035]: I1124 19:14:38.681370 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:39 crc kubenswrapper[5035]: I1124 19:14:39.365431 5035 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.562540 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.562844 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.564269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.564328 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.564342 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.568620 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.685709 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.685797 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.687410 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.687458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:40 crc kubenswrapper[5035]: I1124 19:14:40.687469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:41 crc kubenswrapper[5035]: E1124 19:14:41.158813 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.160908 5035 trace.go:236] Trace[647851953]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 19:14:29.267) (total time: 11893ms): Nov 24 19:14:41 crc kubenswrapper[5035]: Trace[647851953]: ---"Objects listed" error: 11893ms (19:14:41.160) Nov 24 19:14:41 crc kubenswrapper[5035]: Trace[647851953]: [11.893597654s] [11.893597654s] END Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.160935 5035 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.161577 5035 trace.go:236] Trace[696978207]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 19:14:30.232) (total time: 10928ms): Nov 24 19:14:41 crc kubenswrapper[5035]: Trace[696978207]: ---"Objects listed" error: 10928ms (19:14:41.161) Nov 24 19:14:41 crc kubenswrapper[5035]: Trace[696978207]: [10.928666113s] [10.928666113s] END Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.161610 5035 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.165094 5035 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.166459 5035 trace.go:236] Trace[1219582415]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 19:14:26.355) (total time: 14809ms): Nov 24 19:14:41 crc kubenswrapper[5035]: Trace[1219582415]: ---"Objects listed" error: 14809ms (19:14:41.164) Nov 24 19:14:41 crc kubenswrapper[5035]: Trace[1219582415]: [14.809152067s] [14.809152067s] END Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.166541 5035 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 19:14:41 crc kubenswrapper[5035]: E1124 19:14:41.167697 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.203244 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48676->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.203315 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48676->192.168.126.11:17697: read: connection reset by peer" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.203354 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48680->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.203444 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48680->192.168.126.11:17697: read: connection reset by peer" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.206629 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48688->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.206665 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48688->192.168.126.11:17697: read: connection reset by peer" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.214007 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.217505 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.691449 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.694315 5035 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be" exitCode=255 Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.694352 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be"} Nov 24 19:14:41 crc kubenswrapper[5035]: E1124 19:14:41.701851 5035 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 19:14:41 crc kubenswrapper[5035]: I1124 19:14:41.708734 5035 scope.go:117] "RemoveContainer" containerID="44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.124615 5035 apiserver.go:52] "Watching apiserver" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.127432 5035 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.127696 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.127997 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.128111 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.128201 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.128243 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.128127 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.129324 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.129541 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.130396 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.128194 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.130783 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.131402 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.131679 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.132057 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.134915 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.134990 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.135432 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.135442 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.135465 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.171550 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173234 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173327 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173367 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173411 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173445 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173480 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173582 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173616 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173697 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173732 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173795 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173833 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.173866 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.173972 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.174045 5035 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.174062 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:42.674032685 +0000 UTC m=+21.196538982 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.174511 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.174575 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.174647 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:42.674623131 +0000 UTC m=+21.197129478 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.174700 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.175417 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.183942 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.184054 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.195253 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.195313 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.195333 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.195414 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:42.695390303 +0000 UTC m=+21.217896590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.195429 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.196470 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.201102 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.205929 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.205992 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.206012 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.206024 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.206097 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:42.706062391 +0000 UTC m=+21.228568648 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.209636 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.216757 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.229559 5035 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.236365 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.251967 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.263158 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274342 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274389 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274416 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274440 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274464 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274511 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274536 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274559 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274582 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274610 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274638 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274671 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274706 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274733 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274758 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274778 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274800 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274822 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274827 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274843 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274953 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274972 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274990 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275005 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.274975 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275063 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275079 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275098 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275105 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275115 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275182 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275239 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275226 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275262 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275322 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275347 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275431 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275461 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275505 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275525 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275542 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275580 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275602 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275625 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275691 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.275856 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276137 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276179 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276228 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276282 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276407 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276584 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276638 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276650 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276743 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.276771 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277034 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277070 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278822 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277280 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277415 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278878 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277439 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277451 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277536 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277852 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277875 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277930 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.277928 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278247 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278247 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278414 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278430 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278576 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278672 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278752 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.278829 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279245 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279262 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279387 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279791 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279820 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279868 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279895 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279894 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279916 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279942 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279968 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.279988 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280011 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280034 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280069 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280050 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280090 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280222 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280245 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280214 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280361 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280390 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280480 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280444 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280505 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280530 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280554 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280637 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280663 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280725 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280775 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280674 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280811 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280859 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280884 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280910 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280933 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280951 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.280988 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281023 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281123 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281144 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281166 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281188 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281186 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281213 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281314 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281342 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281451 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281468 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281539 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281562 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281697 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281716 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281737 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281758 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281776 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281796 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281819 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281838 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281860 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281884 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281907 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281977 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282094 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282117 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282134 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282158 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282177 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282200 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282229 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282250 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282270 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282287 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282473 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282495 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282579 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282621 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282640 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282724 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282753 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282782 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282812 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282846 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282924 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281312 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281535 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281544 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283191 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281583 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281857 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281865 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.281972 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282849 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282662 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282969 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282972 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283470 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283552 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.282601 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283212 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283686 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283832 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283889 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283930 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283977 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.284101 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.284241 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.284278 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.285112 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.285149 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.285236 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.285748 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.285813 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.283427 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.285989 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.286003 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.286121 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.286144 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.286215 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.286321 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.286538 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.286557 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.287075 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.287465 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.287480 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.287546 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.287770 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.287817 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.287989 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.288085 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:14:42.78806484 +0000 UTC m=+21.310571097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.288153 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.288899 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.288993 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.289224 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.289866 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.288475 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292447 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.288157 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292820 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292848 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292882 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292889 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292921 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292946 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292969 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.292995 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293075 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293103 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293125 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293147 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293171 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293192 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293215 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293238 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293260 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293283 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293325 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293348 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293371 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293394 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293416 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293421 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293439 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293462 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293485 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293507 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293495 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293529 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293508 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293545 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293553 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293656 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293662 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293701 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293722 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293756 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293782 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293805 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293823 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293843 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293883 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293910 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293940 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293960 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293987 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294005 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294059 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294086 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294102 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294117 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294133 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294148 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294190 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294207 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294230 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294252 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294276 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294303 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294320 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294338 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294364 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294380 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294396 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294423 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294440 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294501 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294517 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294537 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294555 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294571 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294586 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294603 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294619 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294635 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294660 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294678 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294696 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294713 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294729 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294745 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294760 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294776 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294883 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295015 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295236 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295256 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295282 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295321 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295334 5035 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295348 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295361 5035 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295372 5035 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295386 5035 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295401 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295413 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295424 5035 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295437 5035 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295449 5035 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295462 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295476 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295488 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295500 5035 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295512 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295525 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295538 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295551 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295593 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295606 5035 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295618 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295632 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295643 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295655 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295666 5035 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295677 5035 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295688 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295701 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295714 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295725 5035 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295737 5035 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295750 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295763 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295772 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295781 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295793 5035 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295806 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295821 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295832 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295843 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295855 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295867 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295878 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295891 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295903 5035 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295915 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295930 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295949 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295965 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295976 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295986 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295996 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296007 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296017 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296027 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296037 5035 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296048 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296059 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296071 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296098 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296109 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296122 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296133 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296144 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296156 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296169 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296190 5035 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296204 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296217 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296228 5035 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296240 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296252 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296264 5035 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296275 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296301 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296313 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296336 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296350 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296361 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296372 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296387 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296400 5035 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296412 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296424 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296436 5035 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296458 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296470 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296482 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296495 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296508 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296520 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296533 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296546 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296572 5035 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296586 5035 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296599 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296611 5035 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296623 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296645 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296653 5035 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296662 5035 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296706 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293919 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294000 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294172 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.293972 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294310 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294575 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294643 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294857 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.294906 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295122 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295303 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295386 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295402 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.295507 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296061 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296068 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296171 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296658 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.296673 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297099 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297223 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297435 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297434 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297447 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297531 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297532 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.297544 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298259 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298318 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298479 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298798 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298843 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298883 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.299442 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.299465 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.299681 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.299852 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.299879 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.299981 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298904 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.298926 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300019 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.299121 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300360 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300440 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300467 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.288738 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.288865 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300596 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300715 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300754 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300744 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.300958 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.301122 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.301165 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.301317 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.301540 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.301939 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.302028 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.302044 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.302094 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.303528 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.303835 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.304592 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.304656 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.305852 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.305879 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.306442 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.308014 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.309802 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.309890 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.309977 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.310276 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.310615 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.310757 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.310813 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.310926 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.310954 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.311384 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.311462 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.311513 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.311510 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.311615 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.311881 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.312243 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.312683 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.313164 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.321805 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.322681 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.332312 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.333901 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.336131 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.341535 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.348511 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.359196 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.368906 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.376339 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.385381 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.393131 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.397942 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398089 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398201 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398326 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398456 5035 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398570 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398671 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398775 5035 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.398951 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399124 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399271 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399396 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399496 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399610 5035 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399714 5035 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399824 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.399926 5035 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400035 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400137 5035 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400240 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400394 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400539 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400647 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400745 5035 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400844 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.400942 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401053 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401154 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401274 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401419 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401530 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401663 5035 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401813 5035 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.401954 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402117 5035 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402233 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402376 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402476 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402622 5035 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402732 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402839 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402926 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403010 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403143 5035 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403239 5035 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403345 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403529 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403624 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403750 5035 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403856 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.403942 5035 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404025 5035 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404108 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404184 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404318 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404410 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404494 5035 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404591 5035 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404688 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404806 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.404904 5035 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.405010 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.405122 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.405222 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.405464 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.409784 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.409891 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.409971 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410046 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410137 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410214 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410316 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410414 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410494 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410577 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410650 5035 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410739 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410836 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.410926 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411196 5035 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411276 5035 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411419 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411522 5035 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411609 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411695 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411778 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411858 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.411940 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.412032 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.402443 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.447661 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 19:14:42 crc kubenswrapper[5035]: W1124 19:14:42.457961 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d985a88e911c3baa5c04485098ae3f185414ce45d0817cce451dd12005d96c3f WatchSource:0}: Error finding container d985a88e911c3baa5c04485098ae3f185414ce45d0817cce451dd12005d96c3f: Status 404 returned error can't find the container with id d985a88e911c3baa5c04485098ae3f185414ce45d0817cce451dd12005d96c3f Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.484828 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 19:14:42 crc kubenswrapper[5035]: W1124 19:14:42.496802 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-7a711f839d800b64bd77e00ddea5883e879e36ebf30de5afee41651618fbe43f WatchSource:0}: Error finding container 7a711f839d800b64bd77e00ddea5883e879e36ebf30de5afee41651618fbe43f: Status 404 returned error can't find the container with id 7a711f839d800b64bd77e00ddea5883e879e36ebf30de5afee41651618fbe43f Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.698989 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.702551 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b"} Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.703599 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.704782 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7a711f839d800b64bd77e00ddea5883e879e36ebf30de5afee41651618fbe43f"} Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.706266 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb"} Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.706404 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d985a88e911c3baa5c04485098ae3f185414ce45d0817cce451dd12005d96c3f"} Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.709301 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc"} Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.709408 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697"} Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.709483 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d2a9ad309374603c03237d147c70c1e8136f77e333aec4a1094364fb7531111c"} Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.714928 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.715088 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.715199 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.715351 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.715218 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.715562 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.715633 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.715735 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:43.715718307 +0000 UTC m=+22.238224564 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.715314 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.715360 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.716215 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.716236 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.715503 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.716176 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:43.71616491 +0000 UTC m=+22.238671167 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.716371 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:43.716353225 +0000 UTC m=+22.238859512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.716398 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:43.716381826 +0000 UTC m=+22.238888123 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.720338 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.731060 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.742409 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.753168 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.765051 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.778166 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.787610 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.798413 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.808360 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.816257 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:42 crc kubenswrapper[5035]: E1124 19:14:42.816402 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:14:43.81637866 +0000 UTC m=+22.338884937 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.819330 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.833241 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.849040 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.863885 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.878818 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.892886 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:42 crc kubenswrapper[5035]: I1124 19:14:42.907612 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:43 crc kubenswrapper[5035]: I1124 19:14:43.199050 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.199176 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:43 crc kubenswrapper[5035]: I1124 19:14:43.724557 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:43 crc kubenswrapper[5035]: I1124 19:14:43.724608 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:43 crc kubenswrapper[5035]: I1124 19:14:43.724632 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:43 crc kubenswrapper[5035]: I1124 19:14:43.724656 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724792 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724802 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724856 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724870 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724882 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724921 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:45.724886255 +0000 UTC m=+24.247392562 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724932 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724811 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724970 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.724959 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:45.724940846 +0000 UTC m=+24.247447153 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.725006 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:45.724995377 +0000 UTC m=+24.247501644 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.725020 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:45.725013338 +0000 UTC m=+24.247519605 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:43 crc kubenswrapper[5035]: I1124 19:14:43.825542 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:43 crc kubenswrapper[5035]: E1124 19:14:43.825770 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:14:45.825737532 +0000 UTC m=+24.348243799 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.200156 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.200217 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:44 crc kubenswrapper[5035]: E1124 19:14:44.200394 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:44 crc kubenswrapper[5035]: E1124 19:14:44.200538 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.206078 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.206586 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.207650 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.208228 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.209150 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.209616 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.210152 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.211069 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.211634 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.212497 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.212937 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.213978 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.214478 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.214940 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.215807 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.216395 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.217523 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.217917 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.218444 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.219547 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.219965 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.220931 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.221345 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.222382 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.222833 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.223394 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.224401 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.224825 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.226098 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.226542 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.227437 5035 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.227533 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.229180 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.230282 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.230889 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.233129 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.234199 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.235773 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.236690 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.237888 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.238405 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.239486 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.240307 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.241556 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.242169 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.243199 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.243922 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.245137 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.245636 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.246502 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.246924 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.247785 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.248361 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 19:14:44 crc kubenswrapper[5035]: I1124 19:14:44.248858 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.199933 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.200197 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.717886 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03"} Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.732196 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.742368 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.742432 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.742472 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.742521 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742588 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742644 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742669 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742684 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742713 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742703 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742749 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742677 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:49.742651454 +0000 UTC m=+28.265157721 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742759 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742799 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:49.742772717 +0000 UTC m=+28.265278984 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742829 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:49.742812148 +0000 UTC m=+28.265318445 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.742856 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:49.742844169 +0000 UTC m=+28.265350466 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.745000 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.756535 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.767515 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.783541 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.799278 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.810480 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.823081 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:45 crc kubenswrapper[5035]: I1124 19:14:45.843487 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:45 crc kubenswrapper[5035]: E1124 19:14:45.843684 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:14:49.843657806 +0000 UTC m=+28.366164063 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:14:46 crc kubenswrapper[5035]: I1124 19:14:46.199889 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:46 crc kubenswrapper[5035]: I1124 19:14:46.199932 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:46 crc kubenswrapper[5035]: E1124 19:14:46.200104 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:46 crc kubenswrapper[5035]: E1124 19:14:46.200165 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.199985 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:47 crc kubenswrapper[5035]: E1124 19:14:47.200133 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.568225 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.570017 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.570178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.570281 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.570466 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.586900 5035 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.587412 5035 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.588551 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.588584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.588594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.588613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.588624 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: E1124 19:14:47.635153 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.641382 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.641431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.641445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.641461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.641471 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: E1124 19:14:47.657089 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.660858 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.660886 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.660894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.660907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.660917 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: E1124 19:14:47.674841 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.680151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.680401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.680478 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.680575 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.680656 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: E1124 19:14:47.696389 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.701233 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.701274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.701283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.701311 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.701342 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: E1124 19:14:47.715202 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: E1124 19:14:47.715393 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.716906 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.716932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.716941 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.716956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.716964 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.720124 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xbc8w"] Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.720835 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-mg7qv"] Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.721109 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-tqwj5"] Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.721208 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.721125 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.721831 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.724341 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725097 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725226 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725243 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725396 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725450 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725517 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725699 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.725721 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.728249 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.740818 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.757365 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.771308 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.783969 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.798859 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.814377 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.818593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.818637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.818647 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.818661 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.818671 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.825095 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.836533 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.850204 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858324 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-cni-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858397 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m5c4\" (UniqueName: \"kubernetes.io/projected/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-kube-api-access-9m5c4\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858539 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-os-release\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858696 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/03570c32-b3fc-4672-9dcc-619560df1077-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858739 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-netns\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858761 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-hostroot\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858787 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-system-cni-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858814 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-cni-multus\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858831 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-conf-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858854 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-kubelet\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858870 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drlbp\" (UniqueName: \"kubernetes.io/projected/03570c32-b3fc-4672-9dcc-619560df1077-kube-api-access-drlbp\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858886 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4g5t\" (UniqueName: \"kubernetes.io/projected/d6b22ba0-c1c5-4d4b-a402-966e27b725ec-kube-api-access-w4g5t\") pod \"node-resolver-tqwj5\" (UID: \"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\") " pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858903 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-daemon-config\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.858924 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-os-release\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859005 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-etc-kubernetes\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859046 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-k8s-cni-cncf-io\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859076 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-cnibin\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859130 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/03570c32-b3fc-4672-9dcc-619560df1077-cni-binary-copy\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859171 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859303 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-system-cni-dir\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859369 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-cni-binary-copy\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859402 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d6b22ba0-c1c5-4d4b-a402-966e27b725ec-hosts-file\") pod \"node-resolver-tqwj5\" (UID: \"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\") " pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859444 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-cnibin\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859467 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-multus-certs\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859490 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-socket-dir-parent\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.859514 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-cni-bin\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.865342 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.877955 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.899223 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.917527 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.921238 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.921280 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.921311 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.921331 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.921340 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:47Z","lastTransitionTime":"2025-11-24T19:14:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.933385 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.946940 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960016 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960143 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-cni-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960196 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m5c4\" (UniqueName: \"kubernetes.io/projected/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-kube-api-access-9m5c4\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960218 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-os-release\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960251 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-netns\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960274 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-hostroot\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960322 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/03570c32-b3fc-4672-9dcc-619560df1077-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960348 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-system-cni-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960361 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-netns\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960369 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-cni-multus\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960414 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-cni-multus\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960425 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-os-release\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-kubelet\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960460 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-hostroot\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960474 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-conf-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960501 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-kubelet\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960496 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-cni-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960502 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drlbp\" (UniqueName: \"kubernetes.io/projected/03570c32-b3fc-4672-9dcc-619560df1077-kube-api-access-drlbp\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960580 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4g5t\" (UniqueName: \"kubernetes.io/projected/d6b22ba0-c1c5-4d4b-a402-966e27b725ec-kube-api-access-w4g5t\") pod \"node-resolver-tqwj5\" (UID: \"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\") " pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960592 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-system-cni-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960600 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-daemon-config\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960543 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-conf-dir\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960618 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-os-release\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960637 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-etc-kubernetes\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960654 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-k8s-cni-cncf-io\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960672 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-cnibin\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960687 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/03570c32-b3fc-4672-9dcc-619560df1077-cni-binary-copy\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960703 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960720 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-system-cni-dir\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960746 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-cnibin\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960761 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-cni-binary-copy\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960779 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d6b22ba0-c1c5-4d4b-a402-966e27b725ec-hosts-file\") pod \"node-resolver-tqwj5\" (UID: \"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\") " pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960795 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-socket-dir-parent\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960809 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-cni-bin\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960824 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-multus-certs\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960872 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-multus-certs\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960865 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-os-release\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960913 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-system-cni-dir\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960936 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-cnibin\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960956 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d6b22ba0-c1c5-4d4b-a402-966e27b725ec-hosts-file\") pod \"node-resolver-tqwj5\" (UID: \"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\") " pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960967 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-socket-dir-parent\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960988 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-var-lib-cni-bin\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.960992 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-etc-kubernetes\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.961008 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-cnibin\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.961036 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-host-run-k8s-cni-cncf-io\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.961319 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/03570c32-b3fc-4672-9dcc-619560df1077-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.961431 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-multus-daemon-config\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.961596 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/03570c32-b3fc-4672-9dcc-619560df1077-cni-binary-copy\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.961658 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-cni-binary-copy\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.961745 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/03570c32-b3fc-4672-9dcc-619560df1077-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.979637 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m5c4\" (UniqueName: \"kubernetes.io/projected/38b324c3-dc7d-4555-ac0d-714a9f1a40cf-kube-api-access-9m5c4\") pod \"multus-mg7qv\" (UID: \"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\") " pod="openshift-multus/multus-mg7qv" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.979639 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.980339 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drlbp\" (UniqueName: \"kubernetes.io/projected/03570c32-b3fc-4672-9dcc-619560df1077-kube-api-access-drlbp\") pod \"multus-additional-cni-plugins-xbc8w\" (UID: \"03570c32-b3fc-4672-9dcc-619560df1077\") " pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.982139 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4g5t\" (UniqueName: \"kubernetes.io/projected/d6b22ba0-c1c5-4d4b-a402-966e27b725ec-kube-api-access-w4g5t\") pod \"node-resolver-tqwj5\" (UID: \"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\") " pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:47 crc kubenswrapper[5035]: I1124 19:14:47.989922 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:47.999986 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:47Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.011522 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.023732 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.023991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.024084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.023826 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.024392 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.024425 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.038008 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mg7qv" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.045372 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.050528 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tqwj5" Nov 24 19:14:48 crc kubenswrapper[5035]: W1124 19:14:48.059299 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03570c32_b3fc_4672_9dcc_619560df1077.slice/crio-8b4f77faa5a05fa4f247a3657343605e70c80cd19facba61b27fddfbbf1f0935 WatchSource:0}: Error finding container 8b4f77faa5a05fa4f247a3657343605e70c80cd19facba61b27fddfbbf1f0935: Status 404 returned error can't find the container with id 8b4f77faa5a05fa4f247a3657343605e70c80cd19facba61b27fddfbbf1f0935 Nov 24 19:14:48 crc kubenswrapper[5035]: W1124 19:14:48.066466 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6b22ba0_c1c5_4d4b_a402_966e27b725ec.slice/crio-41f5306d42c8429180a53ba9b0ee7f1a46b2d36246a99ef2721cf2d15194e827 WatchSource:0}: Error finding container 41f5306d42c8429180a53ba9b0ee7f1a46b2d36246a99ef2721cf2d15194e827: Status 404 returned error can't find the container with id 41f5306d42c8429180a53ba9b0ee7f1a46b2d36246a99ef2721cf2d15194e827 Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.079892 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-nvql4"] Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.080476 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.080857 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dbbzx"] Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.081663 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.082701 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.083740 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.083972 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.084144 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.084277 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.084463 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.084659 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.086097 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.086418 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.086609 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.086822 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.086848 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.097001 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.113786 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.128163 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.132141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.132179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.132191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.132208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.132220 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.147268 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.159511 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.176877 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.191914 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.200062 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.200112 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:48 crc kubenswrapper[5035]: E1124 19:14:48.200177 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:48 crc kubenswrapper[5035]: E1124 19:14:48.200537 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.203028 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.216244 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.228589 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.235180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.235213 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.235221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.235235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.235244 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.240028 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.252873 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263150 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263212 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-proxy-tls\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263229 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-env-overrides\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263242 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-netns\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263274 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-systemd\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263307 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-mcd-auth-proxy-config\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263322 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-slash\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263341 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-ovn-kubernetes\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263361 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-config\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263378 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovn-node-metrics-cert\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263392 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-kubelet\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263405 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-log-socket\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263419 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-rootfs\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263432 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-var-lib-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263448 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-netd\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4qfz\" (UniqueName: \"kubernetes.io/projected/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-kube-api-access-x4qfz\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263476 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-etc-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263491 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-ovn\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263506 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-node-log\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263560 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbkht\" (UniqueName: \"kubernetes.io/projected/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-kube-api-access-lbkht\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263627 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-bin\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263688 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-systemd-units\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263787 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-script-lib\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.263831 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.265853 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.278304 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.305542 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.334068 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.337936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.337973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.337984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.338002 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.338012 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.352955 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364568 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-script-lib\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364609 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364629 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-proxy-tls\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364671 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-env-overrides\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364686 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-systemd\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364706 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-mcd-auth-proxy-config\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364722 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-netns\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364737 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-ovn-kubernetes\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364750 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-config\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364764 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovn-node-metrics-cert\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364772 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364815 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-kubelet\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364780 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-kubelet\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364844 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-netns\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-slash\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364874 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-slash\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364888 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-log-socket\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364912 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-rootfs\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364930 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-var-lib-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364954 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4qfz\" (UniqueName: \"kubernetes.io/projected/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-kube-api-access-x4qfz\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364971 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-etc-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364989 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-netd\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365006 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-node-log\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365022 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbkht\" (UniqueName: \"kubernetes.io/projected/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-kube-api-access-lbkht\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365060 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-ovn\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365082 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-bin\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365100 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-systemd-units\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365170 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-systemd-units\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365194 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-systemd\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365227 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-env-overrides\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365300 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-ovn-kubernetes\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365331 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-etc-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.364840 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365353 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-script-lib\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365361 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-log-socket\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365404 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-var-lib-openvswitch\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365380 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-rootfs\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365542 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-config\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365553 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-netd\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365587 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-node-log\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365605 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-ovn\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365620 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-bin\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.365841 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-mcd-auth-proxy-config\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.368105 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.368381 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovn-node-metrics-cert\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.368523 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-proxy-tls\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.380134 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.380348 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbkht\" (UniqueName: \"kubernetes.io/projected/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-kube-api-access-lbkht\") pod \"ovnkube-node-dbbzx\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.383442 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4qfz\" (UniqueName: \"kubernetes.io/projected/11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361-kube-api-access-x4qfz\") pod \"machine-config-daemon-nvql4\" (UID: \"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\") " pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.391584 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.400971 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.412906 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.413979 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:14:48 crc kubenswrapper[5035]: W1124 19:14:48.424027 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11fbb8c9_66d2_4fdd_bb91_b4dfb6ea9361.slice/crio-3f7b53d9ce2cadf89a0fa514e2f1c3c9efcefed56122302dfe638234666964f9 WatchSource:0}: Error finding container 3f7b53d9ce2cadf89a0fa514e2f1c3c9efcefed56122302dfe638234666964f9: Status 404 returned error can't find the container with id 3f7b53d9ce2cadf89a0fa514e2f1c3c9efcefed56122302dfe638234666964f9 Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.424758 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.436887 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.440444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.440486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.440497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.440511 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.440522 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.443962 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: W1124 19:14:48.446791 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d9cb31b_5f2a_4594_ad12_718b6e99d15f.slice/crio-2c580bb5b6b4c531b89e2ac918b89057fd8714911f549d6138246fbbddd1d910 WatchSource:0}: Error finding container 2c580bb5b6b4c531b89e2ac918b89057fd8714911f549d6138246fbbddd1d910: Status 404 returned error can't find the container with id 2c580bb5b6b4c531b89e2ac918b89057fd8714911f549d6138246fbbddd1d910 Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.459978 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.543521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.543561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.543570 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.543585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.543594 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.645975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.646010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.646020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.646035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.646043 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.727368 5035 generic.go:334] "Generic (PLEG): container finished" podID="03570c32-b3fc-4672-9dcc-619560df1077" containerID="8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b" exitCode=0 Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.727437 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerDied","Data":"8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.727463 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerStarted","Data":"8b4f77faa5a05fa4f247a3657343605e70c80cd19facba61b27fddfbbf1f0935"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.729716 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerStarted","Data":"6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.729785 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerStarted","Data":"0ca9e4e66711452237a71a5e459938e8441d1dfb835254d2ff2055bbd437e76f"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.732389 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3" exitCode=0 Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.732437 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.732526 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"2c580bb5b6b4c531b89e2ac918b89057fd8714911f549d6138246fbbddd1d910"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.734200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.734252 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"3f7b53d9ce2cadf89a0fa514e2f1c3c9efcefed56122302dfe638234666964f9"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.736087 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tqwj5" event={"ID":"d6b22ba0-c1c5-4d4b-a402-966e27b725ec","Type":"ContainerStarted","Data":"e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.736133 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tqwj5" event={"ID":"d6b22ba0-c1c5-4d4b-a402-966e27b725ec","Type":"ContainerStarted","Data":"41f5306d42c8429180a53ba9b0ee7f1a46b2d36246a99ef2721cf2d15194e827"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.745626 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.752469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.752517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.752533 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.752555 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.752571 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.762900 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.774854 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.791009 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.819912 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.838746 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.853581 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.855710 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.855753 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.855765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.855785 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.855797 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.875104 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.888682 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.904183 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.916887 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.931007 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.942576 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.958606 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.958902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.958912 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.958926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.958935 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:48Z","lastTransitionTime":"2025-11-24T19:14:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.968245 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.980626 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:48 crc kubenswrapper[5035]: I1124 19:14:48.995517 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:48Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.012176 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.023370 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.036914 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.053629 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.060881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.060918 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.060930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.060947 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.060959 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.063657 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.073501 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.089691 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.104487 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.115882 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.149063 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.163544 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.163581 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.163590 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.163604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.163613 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.199327 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.199527 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.265171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.265208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.265217 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.265233 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.265243 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.367256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.367320 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.367335 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.367355 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.367368 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.469757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.469801 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.469811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.469828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.469837 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.572137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.572174 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.572183 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.572199 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.572210 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.674624 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.674661 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.674671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.674686 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.674698 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.740639 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerStarted","Data":"85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.744731 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.744791 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.744805 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.744817 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.744829 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.746686 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.759096 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.773637 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.777682 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.777726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.777736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.777752 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.777765 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.782342 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.782392 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.782441 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.782468 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.782555 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.782592 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.782608 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.782667 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:57.782647103 +0000 UTC m=+36.305153400 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.782555 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.782712 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:57.782704395 +0000 UTC m=+36.305210752 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.782916 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.783007 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:57.782984183 +0000 UTC m=+36.305490520 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.783558 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.783584 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.783596 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.783640 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:14:57.783628049 +0000 UTC m=+36.306134526 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.788783 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.802384 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.814036 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.825638 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.844733 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.855721 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.867344 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.879631 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.879666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.879675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.879688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.879697 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.880653 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.882778 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:49 crc kubenswrapper[5035]: E1124 19:14:49.882934 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:14:57.882902405 +0000 UTC m=+36.405408672 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.891082 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.907968 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.927041 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.942565 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.953779 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.967326 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.981201 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.982434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.982469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.982478 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.982492 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:49 crc kubenswrapper[5035]: I1124 19:14:49.982504 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:49Z","lastTransitionTime":"2025-11-24T19:14:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.002148 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.022070 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.055358 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.080587 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.085215 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.085250 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.085261 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.085275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.085301 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.102835 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.113274 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.126010 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.154107 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.187720 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.187765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.187775 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.187790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.187801 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.195148 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.199667 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:50 crc kubenswrapper[5035]: E1124 19:14:50.200066 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.199725 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:50 crc kubenswrapper[5035]: E1124 19:14:50.200358 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.289990 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.290040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.290053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.290071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.290083 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.392924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.392965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.392975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.392988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.393000 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.496400 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.496473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.496494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.496521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.496539 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.599489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.599552 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.599571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.599595 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.599612 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.702784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.703038 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.703049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.703066 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.703079 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.750524 5035 generic.go:334] "Generic (PLEG): container finished" podID="03570c32-b3fc-4672-9dcc-619560df1077" containerID="85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45" exitCode=0 Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.750583 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerDied","Data":"85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.761190 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.773990 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.791542 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.801765 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.808332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.808360 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.808372 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.808388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.808399 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.818974 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.830146 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.842725 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.862524 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.895099 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.907783 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.910857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.910879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.910887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.910900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.910909 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:50Z","lastTransitionTime":"2025-11-24T19:14:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.928558 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.942724 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.960721 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:50 crc kubenswrapper[5035]: I1124 19:14:50.973140 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:50Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.013186 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.013232 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.013251 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.013273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.013319 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.115998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.116032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.116040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.116054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.116065 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.199576 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:51 crc kubenswrapper[5035]: E1124 19:14:51.199818 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.219109 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.219141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.219150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.219162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.219171 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.321930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.321966 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.321974 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.321986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.321994 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.426252 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.426609 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.426626 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.426650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.426668 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.430258 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-fjhbz"] Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.430959 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.434690 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.435153 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.435489 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.435530 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.457348 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.475476 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.490019 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.496562 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56bc7a88-542e-4c82-86bb-8a4537e850b9-serviceca\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.496609 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqwmg\" (UniqueName: \"kubernetes.io/projected/56bc7a88-542e-4c82-86bb-8a4537e850b9-kube-api-access-cqwmg\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.496664 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56bc7a88-542e-4c82-86bb-8a4537e850b9-host\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.505336 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.516647 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.529272 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.529377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.529405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.529439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.529463 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.548481 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.563471 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.583417 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.597034 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56bc7a88-542e-4c82-86bb-8a4537e850b9-host\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.597088 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56bc7a88-542e-4c82-86bb-8a4537e850b9-serviceca\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.597112 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqwmg\" (UniqueName: \"kubernetes.io/projected/56bc7a88-542e-4c82-86bb-8a4537e850b9-kube-api-access-cqwmg\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.597261 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56bc7a88-542e-4c82-86bb-8a4537e850b9-host\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.599240 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56bc7a88-542e-4c82-86bb-8a4537e850b9-serviceca\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.615465 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.617093 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqwmg\" (UniqueName: \"kubernetes.io/projected/56bc7a88-542e-4c82-86bb-8a4537e850b9-kube-api-access-cqwmg\") pod \"node-ca-fjhbz\" (UID: \"56bc7a88-542e-4c82-86bb-8a4537e850b9\") " pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.633454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.633519 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.633545 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.633579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.633601 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.633753 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.646018 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.657506 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.671257 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.685060 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.736124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.736161 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.736170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.736185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.736198 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.746401 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fjhbz" Nov 24 19:14:51 crc kubenswrapper[5035]: W1124 19:14:51.760122 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56bc7a88_542e_4c82_86bb_8a4537e850b9.slice/crio-1906b699bd76bebb49492d01cd955eadd3b3d1ce484d8a058943826cea03218b WatchSource:0}: Error finding container 1906b699bd76bebb49492d01cd955eadd3b3d1ce484d8a058943826cea03218b: Status 404 returned error can't find the container with id 1906b699bd76bebb49492d01cd955eadd3b3d1ce484d8a058943826cea03218b Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.764919 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fjhbz" event={"ID":"56bc7a88-542e-4c82-86bb-8a4537e850b9","Type":"ContainerStarted","Data":"1906b699bd76bebb49492d01cd955eadd3b3d1ce484d8a058943826cea03218b"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.766675 5035 generic.go:334] "Generic (PLEG): container finished" podID="03570c32-b3fc-4672-9dcc-619560df1077" containerID="48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369" exitCode=0 Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.766714 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerDied","Data":"48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.781492 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.804761 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.818686 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.839437 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.840307 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.840349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.840364 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.840383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.840393 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.848759 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.863544 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.873779 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.883920 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.898030 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.916961 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.932231 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.942193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.942242 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.942260 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.942282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.942334 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:51Z","lastTransitionTime":"2025-11-24T19:14:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.945099 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.956739 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:51 crc kubenswrapper[5035]: I1124 19:14:51.968347 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:51Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.045369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.045580 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.045589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.045602 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.045614 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.148878 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.148931 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.148949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.148976 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.148997 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.199123 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.199212 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:52 crc kubenswrapper[5035]: E1124 19:14:52.199355 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:52 crc kubenswrapper[5035]: E1124 19:14:52.199510 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.228497 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.245779 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.252057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.252116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.252135 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.252159 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.252177 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.263391 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.284165 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.304345 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.325887 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.344609 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.353895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.353941 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.353951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.353967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.353977 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.365007 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.376942 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.394125 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.413983 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.428400 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.456279 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.456354 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.456366 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.456381 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.456393 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.459616 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.492188 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.558527 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.558573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.558588 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.558610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.558627 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.661474 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.661517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.661529 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.661547 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.661575 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.764928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.764972 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.764983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.765000 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.765012 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.775238 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.777389 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fjhbz" event={"ID":"56bc7a88-542e-4c82-86bb-8a4537e850b9","Type":"ContainerStarted","Data":"26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.781793 5035 generic.go:334] "Generic (PLEG): container finished" podID="03570c32-b3fc-4672-9dcc-619560df1077" containerID="34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66" exitCode=0 Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.781852 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerDied","Data":"34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.800162 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.819811 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.840382 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.857525 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.866862 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.866892 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.866900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.866913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.866921 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.878304 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.894579 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.906615 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.918563 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.934813 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.946993 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.960208 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.969703 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.969745 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.969758 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.969775 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.969788 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:52Z","lastTransitionTime":"2025-11-24T19:14:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:52 crc kubenswrapper[5035]: I1124 19:14:52.973327 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.017197 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.050132 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.071593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.071649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.071665 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.071687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.071703 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.093366 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.133068 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.171551 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.174159 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.174191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.174200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.174213 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.174222 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.200059 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:53 crc kubenswrapper[5035]: E1124 19:14:53.200192 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.211794 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.252064 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.276680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.276726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.276745 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.276765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.276779 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.291649 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.331682 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.372811 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.380354 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.380416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.380433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.380728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.380774 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.417767 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.462758 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.484559 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.484627 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.484646 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.484676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.484696 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.499181 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.534547 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.587571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.587640 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.587666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.587698 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.587720 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.590499 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.620808 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.691268 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.691375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.691399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.691432 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.691454 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.791894 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerStarted","Data":"5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.793396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.793460 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.793482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.793507 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.793525 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.813912 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.837474 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.855150 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.869720 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.889802 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.896025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.896064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.896078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.896101 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.896116 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.910404 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.928597 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.954317 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.978634 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:53Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.999866 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.999914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:53 crc kubenswrapper[5035]: I1124 19:14:53.999927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:53.999946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:53.999958 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:53Z","lastTransitionTime":"2025-11-24T19:14:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.015419 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.055146 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.102203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.102261 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.102280 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.102325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.102343 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.108624 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.133729 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.173961 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.199549 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.199605 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:54 crc kubenswrapper[5035]: E1124 19:14:54.199745 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:54 crc kubenswrapper[5035]: E1124 19:14:54.199848 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.204767 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.204812 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.204831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.204854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.204870 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.307982 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.308515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.308540 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.308570 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.308593 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.411451 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.411480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.411494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.411514 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.411529 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.515527 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.515583 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.515596 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.515621 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.515636 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.618253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.618330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.618342 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.618369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.618383 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.721760 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.721849 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.721876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.721906 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.721928 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.799652 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.800691 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.800752 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.807226 5035 generic.go:334] "Generic (PLEG): container finished" podID="03570c32-b3fc-4672-9dcc-619560df1077" containerID="5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af" exitCode=0 Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.807273 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerDied","Data":"5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.819097 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.825559 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.825599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.825623 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.825640 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.825653 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.869974 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.881266 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.882177 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.885559 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.925083 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.930930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.930969 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.930978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.930991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.931000 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:54Z","lastTransitionTime":"2025-11-24T19:14:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.945804 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.966096 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.982312 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:54 crc kubenswrapper[5035]: I1124 19:14:54.991967 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.000567 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:54Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.011449 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.023201 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.033390 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.033426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.033441 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.033488 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.033506 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.034045 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.046753 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.065447 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.081104 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.090945 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.100191 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.110319 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.122393 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.132961 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.136565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.136624 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.136644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.136669 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.136687 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.144856 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.155920 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.164997 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.172496 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.183755 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.199544 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:55 crc kubenswrapper[5035]: E1124 19:14:55.199655 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.211125 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.239875 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.240165 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.240181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.240203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.240216 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.254778 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.292757 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.343251 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.343647 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.343792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.343827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.343847 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.446363 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.446397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.446406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.446419 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.446429 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.548960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.549408 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.549656 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.549874 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.550076 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.653050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.653122 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.653146 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.653176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.653202 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.755771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.755843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.755860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.755885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.755904 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.817897 5035 generic.go:334] "Generic (PLEG): container finished" podID="03570c32-b3fc-4672-9dcc-619560df1077" containerID="c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4" exitCode=0 Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.817984 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerDied","Data":"c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.818107 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.844020 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.858965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.859263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.859426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.859563 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.859683 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.868127 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.885268 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.902637 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.915746 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.928044 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.941381 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.954495 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.963671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.963732 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.963756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.963790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.963812 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:55Z","lastTransitionTime":"2025-11-24T19:14:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.967727 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:55 crc kubenswrapper[5035]: I1124 19:14:55.993398 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:55Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.033347 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.054737 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.067438 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.068531 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.068551 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.068559 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.068572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.068581 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.080302 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.170780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.170810 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.170818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.170831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.170839 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.199169 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:56 crc kubenswrapper[5035]: E1124 19:14:56.199263 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.199175 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:56 crc kubenswrapper[5035]: E1124 19:14:56.199482 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.273780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.273846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.273869 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.273897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.273916 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.376748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.376886 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.376913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.376945 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.376965 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.479864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.479910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.479929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.479951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.479969 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.583340 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.583386 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.583405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.583428 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.583445 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.687332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.687409 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.687431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.687458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.687483 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.700726 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.719522 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.740221 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.762712 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.783028 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.790512 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.790567 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.790589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.790619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.790642 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.805172 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.823894 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.827103 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" event={"ID":"03570c32-b3fc-4672-9dcc-619560df1077","Type":"ContainerStarted","Data":"a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.827161 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.842202 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.862102 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.890911 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.892992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.893042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.893061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.893083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.893098 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.908172 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.928439 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.944117 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.956017 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.967261 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.979254 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.992183 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:56Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.995763 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.995802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.995811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.995827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:56 crc kubenswrapper[5035]: I1124 19:14:56.995836 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:56Z","lastTransitionTime":"2025-11-24T19:14:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.004129 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.015417 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.031731 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.044870 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.060491 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.076880 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.092364 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.098438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.098468 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.098514 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.098531 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.098543 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.105475 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.120860 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.135081 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.144647 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.156204 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.199390 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.199511 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.200735 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.200769 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.200782 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.200795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.200805 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.303479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.303524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.303537 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.303556 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.303568 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.405883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.405925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.405935 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.405952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.405961 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.508597 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.508636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.508649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.508665 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.508676 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.611343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.611378 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.611389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.611403 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.611412 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.714567 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.714617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.714629 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.714649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.714663 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.817639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.817683 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.817699 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.817722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.817739 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.834747 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/0.log" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.839737 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815" exitCode=1 Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.839791 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.840848 5035 scope.go:117] "RemoveContainer" containerID="c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.859653 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.859815 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.859971 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860067 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:13.860033571 +0000 UTC m=+52.382539888 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860254 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860413 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:13.860379341 +0000 UTC m=+52.382885658 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.860481 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.860588 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860625 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860643 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860655 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860787 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860827 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.860851 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.861163 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:13.861129321 +0000 UTC m=+52.383635618 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.861226 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:13.861207513 +0000 UTC m=+52.383713870 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.863874 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.885933 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.907923 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.920249 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.920769 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.924201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.924252 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.924268 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.926270 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.952571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.952620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.952636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.952662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.952680 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.961825 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.961974 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:15:13.961945068 +0000 UTC m=+52.484451355 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.965058 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.971719 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.975229 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.975270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.975280 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.975539 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.975556 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.983744 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: E1124 19:14:57.992278 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.996239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.996875 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.996916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.996941 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.996954 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:57Z","lastTransitionTime":"2025-11-24T19:14:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:57 crc kubenswrapper[5035]: I1124 19:14:57.998277 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:57Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.013799 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: E1124 19:14:58.014119 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.017378 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.017408 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.017418 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.017431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.017443 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.028227 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: E1124 19:14:58.030328 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.036367 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.036396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.036407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.036424 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.036436 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.040163 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: E1124 19:14:58.053376 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: E1124 19:14:58.053531 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.055408 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.055439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.055451 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.055469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.055483 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.056422 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.071128 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.084046 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.093533 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.158797 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.158837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.158846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.158861 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.158870 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.199367 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:14:58 crc kubenswrapper[5035]: E1124 19:14:58.199483 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.199495 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:14:58 crc kubenswrapper[5035]: E1124 19:14:58.199665 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.261094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.261128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.261136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.261149 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.261158 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.363799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.363854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.363865 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.363890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.363904 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.467111 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.467185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.467201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.467223 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.467237 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.569962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.570330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.570348 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.570366 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.570379 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.672916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.672982 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.672997 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.673037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.673049 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.780011 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.780963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.781020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.781054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.781072 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.844790 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/0.log" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.847598 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.847728 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.862032 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.879205 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.884480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.884535 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.884551 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.884576 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.884592 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.897591 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.915463 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.929471 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.948598 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.966237 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.977914 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.987875 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.987915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.987926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.987944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.987958 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:58Z","lastTransitionTime":"2025-11-24T19:14:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:58 crc kubenswrapper[5035]: I1124 19:14:58.995068 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.015098 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.032441 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.048721 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.072425 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.087730 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.090360 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.090397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.090406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.090421 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.090430 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.192871 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.192962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.192980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.193005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.193022 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.199528 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:14:59 crc kubenswrapper[5035]: E1124 19:14:59.199704 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.296069 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.296131 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.296154 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.296215 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.296241 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.400004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.400064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.400120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.400144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.400160 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.433236 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf"] Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.433924 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.436694 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.438578 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.457649 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.472114 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.476816 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101c186b-17e8-4a94-be8e-ae2b58a3e18c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.476880 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101c186b-17e8-4a94-be8e-ae2b58a3e18c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.476966 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5zc4\" (UniqueName: \"kubernetes.io/projected/101c186b-17e8-4a94-be8e-ae2b58a3e18c-kube-api-access-g5zc4\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.477029 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101c186b-17e8-4a94-be8e-ae2b58a3e18c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.486633 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.502734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.502781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.502792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.502810 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.502823 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.515388 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.534006 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.553718 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.570497 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.577710 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101c186b-17e8-4a94-be8e-ae2b58a3e18c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.577763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101c186b-17e8-4a94-be8e-ae2b58a3e18c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.577811 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5zc4\" (UniqueName: \"kubernetes.io/projected/101c186b-17e8-4a94-be8e-ae2b58a3e18c-kube-api-access-g5zc4\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.577853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101c186b-17e8-4a94-be8e-ae2b58a3e18c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.578728 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101c186b-17e8-4a94-be8e-ae2b58a3e18c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.579064 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101c186b-17e8-4a94-be8e-ae2b58a3e18c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.586489 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.587550 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101c186b-17e8-4a94-be8e-ae2b58a3e18c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.604994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.605115 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.605137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.605164 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.605183 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.605219 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.610161 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5zc4\" (UniqueName: \"kubernetes.io/projected/101c186b-17e8-4a94-be8e-ae2b58a3e18c-kube-api-access-g5zc4\") pod \"ovnkube-control-plane-749d76644c-pqsqf\" (UID: \"101c186b-17e8-4a94-be8e-ae2b58a3e18c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.618178 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.628647 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.648222 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.657753 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.666679 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.679276 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.708129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.708176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.708190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.708210 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.708222 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.753581 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" Nov 24 19:14:59 crc kubenswrapper[5035]: W1124 19:14:59.773996 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod101c186b_17e8_4a94_be8e_ae2b58a3e18c.slice/crio-6a460b5ba8f9f143dd2182932095a7a06e00a04c312ac413cb5d06480b815092 WatchSource:0}: Error finding container 6a460b5ba8f9f143dd2182932095a7a06e00a04c312ac413cb5d06480b815092: Status 404 returned error can't find the container with id 6a460b5ba8f9f143dd2182932095a7a06e00a04c312ac413cb5d06480b815092 Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.810719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.810777 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.810794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.810818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.810835 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.868915 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/1.log" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.873180 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/0.log" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.889621 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b" exitCode=1 Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.889695 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.889742 5035 scope.go:117] "RemoveContainer" containerID="c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.890403 5035 scope.go:117] "RemoveContainer" containerID="9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b" Nov 24 19:14:59 crc kubenswrapper[5035]: E1124 19:14:59.890530 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.891380 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" event={"ID":"101c186b-17e8-4a94-be8e-ae2b58a3e18c","Type":"ContainerStarted","Data":"6a460b5ba8f9f143dd2182932095a7a06e00a04c312ac413cb5d06480b815092"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.907768 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.912636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.912674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.912686 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.912704 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.912715 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:14:59Z","lastTransitionTime":"2025-11-24T19:14:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.922512 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.934390 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.946751 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.958550 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.973442 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:14:59 crc kubenswrapper[5035]: I1124 19:14:59.987407 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:59Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.002808 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.012466 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.016041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.016120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.016134 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.016151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.016187 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.025950 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.037474 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.045669 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.057631 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.074973 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.085387 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.118590 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.118625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.118638 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.118654 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.118667 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.201447 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:00 crc kubenswrapper[5035]: E1124 19:15:00.201555 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.201852 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:00 crc kubenswrapper[5035]: E1124 19:15:00.202127 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.222022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.222059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.222070 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.222085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.222096 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.324129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.324188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.324205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.324230 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.324250 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.427843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.427900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.427922 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.427951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.427972 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.530571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.530613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.530627 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.530646 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.530659 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.634476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.634531 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.634553 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.634581 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.634603 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.737394 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.737754 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.737938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.738088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.738275 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.841237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.841592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.841638 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.841667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.841685 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.897019 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/1.log" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.902800 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" event={"ID":"101c186b-17e8-4a94-be8e-ae2b58a3e18c","Type":"ContainerStarted","Data":"33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.902837 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" event={"ID":"101c186b-17e8-4a94-be8e-ae2b58a3e18c","Type":"ContainerStarted","Data":"115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.917140 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.932210 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.944126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.944188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.944210 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.944239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.944260 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:00Z","lastTransitionTime":"2025-11-24T19:15:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.948953 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.960913 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.970080 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.980898 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:00 crc kubenswrapper[5035]: I1124 19:15:00.993721 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.009123 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.021219 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.037286 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.046438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.046492 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.046511 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.046535 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.046554 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.064109 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.086065 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.104234 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.122629 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.139015 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.148795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.148836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.148850 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.148870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.148884 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.199743 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:01 crc kubenswrapper[5035]: E1124 19:15:01.199850 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.251685 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.251755 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.251783 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.251815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.251839 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.354392 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.354428 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.354437 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.354454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.354465 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.376227 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-l2pnt"] Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.377056 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:01 crc kubenswrapper[5035]: E1124 19:15:01.377170 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.396938 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-996cs\" (UniqueName: \"kubernetes.io/projected/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-kube-api-access-996cs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.397028 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.404640 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.416212 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.425457 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.437254 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.448821 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.457039 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.457089 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.457100 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.457117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.457128 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.474026 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.483084 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.498446 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-996cs\" (UniqueName: \"kubernetes.io/projected/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-kube-api-access-996cs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.498502 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:01 crc kubenswrapper[5035]: E1124 19:15:01.498647 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:01 crc kubenswrapper[5035]: E1124 19:15:01.498703 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:01.998685384 +0000 UTC m=+40.521191651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.508495 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.526251 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.532117 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-996cs\" (UniqueName: \"kubernetes.io/projected/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-kube-api-access-996cs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.541651 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.559733 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.559784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.559802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.559826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.559873 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.561223 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.580950 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.599478 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.622645 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.637571 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.653057 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:01Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.662952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.663141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.663332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.663475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.663641 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.766225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.766282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.766365 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.766399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.766419 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.869858 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.869904 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.869920 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.869944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.869963 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.973572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.973633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.973650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.973672 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:01 crc kubenswrapper[5035]: I1124 19:15:01.973689 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:01Z","lastTransitionTime":"2025-11-24T19:15:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.003888 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:02 crc kubenswrapper[5035]: E1124 19:15:02.004185 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:02 crc kubenswrapper[5035]: E1124 19:15:02.004339 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:03.00427844 +0000 UTC m=+41.526784737 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.077436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.077492 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.077507 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.077527 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.077539 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.180928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.180988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.181005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.181028 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.181045 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.199147 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.199181 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:02 crc kubenswrapper[5035]: E1124 19:15:02.199545 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:02 crc kubenswrapper[5035]: E1124 19:15:02.199803 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.230505 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.248108 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.262921 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.283275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.283780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.283962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.283388 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.284381 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.284496 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.303732 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.317927 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.339233 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.355805 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.370472 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.388684 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.388742 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.388762 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.388787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.388804 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.389525 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.406134 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.424004 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.439495 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.458917 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.478444 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.492887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.492939 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.492950 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.492969 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.492983 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.497150 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.596137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.596202 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.596215 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.596237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.596253 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.698536 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.698622 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.698635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.698653 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.698666 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.801686 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.801734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.801743 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.801764 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.801779 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.903903 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.903951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.903962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.903978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:02 crc kubenswrapper[5035]: I1124 19:15:02.903991 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:02Z","lastTransitionTime":"2025-11-24T19:15:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.006426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.006480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.006495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.006515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.006532 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.017589 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:03 crc kubenswrapper[5035]: E1124 19:15:03.017734 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:03 crc kubenswrapper[5035]: E1124 19:15:03.017814 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:05.017792865 +0000 UTC m=+43.540299132 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.109501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.109548 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.109560 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.109579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.109590 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.199500 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.199598 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:03 crc kubenswrapper[5035]: E1124 19:15:03.199655 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:03 crc kubenswrapper[5035]: E1124 19:15:03.199822 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.212211 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.212269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.212337 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.212370 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.212392 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.315511 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.315562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.315576 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.315983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.316047 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.417772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.417839 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.417864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.417894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.417915 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.522743 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.522792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.522804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.522822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.522834 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.625836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.625920 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.625936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.625966 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.625985 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.729403 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.729474 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.729495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.729524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.729545 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.832545 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.832609 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.832626 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.832650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.832668 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.935992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.936084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.936112 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.936145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:03 crc kubenswrapper[5035]: I1124 19:15:03.936168 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:03Z","lastTransitionTime":"2025-11-24T19:15:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.039051 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.039144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.039173 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.039203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.039224 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.142651 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.142697 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.142713 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.142737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.142755 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.199692 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.199720 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:04 crc kubenswrapper[5035]: E1124 19:15:04.199921 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:04 crc kubenswrapper[5035]: E1124 19:15:04.199978 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.245637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.245674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.245684 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.245868 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.245880 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.348473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.348594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.348603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.348616 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.348626 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.451728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.451790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.451812 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.451837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.451853 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.554878 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.554925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.554936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.554957 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.554971 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.657806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.657872 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.657890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.657914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.657932 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.761237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.761283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.761310 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.761328 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.761340 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.864359 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.864399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.864413 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.864436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.864448 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.967124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.967168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.967188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.967206 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:04 crc kubenswrapper[5035]: I1124 19:15:04.967219 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:04Z","lastTransitionTime":"2025-11-24T19:15:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.042876 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:05 crc kubenswrapper[5035]: E1124 19:15:05.043115 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:05 crc kubenswrapper[5035]: E1124 19:15:05.043217 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:09.043187462 +0000 UTC m=+47.565693719 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.069750 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.069808 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.069823 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.069853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.069887 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.172389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.172422 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.172432 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.172446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.172458 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.199941 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.200098 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:05 crc kubenswrapper[5035]: E1124 19:15:05.200326 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:05 crc kubenswrapper[5035]: E1124 19:15:05.200900 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.274848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.274973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.274984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.275000 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.275010 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.377960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.378009 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.378018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.378030 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.378039 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.481053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.481130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.481152 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.481181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.481203 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.584149 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.584183 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.584191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.584205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.584214 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.687221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.687273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.687310 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.687333 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.687350 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.790243 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.790326 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.790347 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.790377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.790399 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.893699 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.893757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.893773 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.893796 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.893810 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.997185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.997245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.997262 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.997314 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:05 crc kubenswrapper[5035]: I1124 19:15:05.997333 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:05Z","lastTransitionTime":"2025-11-24T19:15:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.099748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.099822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.099841 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.099866 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.099884 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.199723 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.199735 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:06 crc kubenswrapper[5035]: E1124 19:15:06.199980 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:06 crc kubenswrapper[5035]: E1124 19:15:06.200097 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.202794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.202863 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.202888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.202917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.202941 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.306269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.306370 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.306389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.306415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.306433 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.410903 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.411057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.411088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.411157 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.411180 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.515428 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.515824 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.515943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.516071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.516159 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.618815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.618879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.618898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.618924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.618941 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.722447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.722515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.722534 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.722561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.722579 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.825856 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.826534 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.826577 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.826607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.826625 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.929804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.929845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.929858 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.929873 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:06 crc kubenswrapper[5035]: I1124 19:15:06.929888 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:06Z","lastTransitionTime":"2025-11-24T19:15:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.032005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.032054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.032071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.032094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.032111 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.135491 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.135566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.135588 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.135618 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.135639 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.200004 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.200015 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:07 crc kubenswrapper[5035]: E1124 19:15:07.200223 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:07 crc kubenswrapper[5035]: E1124 19:15:07.200439 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.238929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.238983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.238999 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.239021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.239038 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.341879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.341922 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.341932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.341948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.341959 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.445524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.445591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.445610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.445634 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.445655 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.548001 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.548040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.548049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.548064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.548074 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.651139 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.651208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.651228 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.651250 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.651268 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.754107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.754181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.754204 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.754233 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.754255 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.857190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.857244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.857263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.857286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.857329 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.959371 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.959402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.959411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.959424 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:07 crc kubenswrapper[5035]: I1124 19:15:07.959432 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:07Z","lastTransitionTime":"2025-11-24T19:15:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.062601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.062662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.062678 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.062702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.062719 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.166469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.166527 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.166544 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.166569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.166586 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.199605 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.199759 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.200053 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.200122 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.269579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.269632 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.269660 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.269690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.269713 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.372848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.372898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.372916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.372942 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.372960 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.420090 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.420147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.420167 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.420188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.420202 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.445490 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:08Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.451473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.451601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.451619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.451642 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.451691 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.469682 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:08Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.473284 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.473316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.473324 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.473336 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.473374 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.486760 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:08Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.491124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.491162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.491175 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.491194 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.491206 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.507313 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:08Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.512497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.512554 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.512568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.512586 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.513009 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.529164 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:08Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:08 crc kubenswrapper[5035]: E1124 19:15:08.529661 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.532200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.532239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.532253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.532271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.532284 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.634692 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.634741 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.634752 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.634770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.634782 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.737371 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.737438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.737466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.737498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.737523 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.840887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.840936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.840948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.840965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.840978 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.943776 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.943840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.943860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.943883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:08 crc kubenswrapper[5035]: I1124 19:15:08.943900 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:08Z","lastTransitionTime":"2025-11-24T19:15:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.046927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.046965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.046996 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.047014 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.047025 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.089108 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:09 crc kubenswrapper[5035]: E1124 19:15:09.089366 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:09 crc kubenswrapper[5035]: E1124 19:15:09.089473 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:17.089444039 +0000 UTC m=+55.611950326 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.150496 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.150562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.150574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.150588 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.150598 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.199512 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.199564 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:09 crc kubenswrapper[5035]: E1124 19:15:09.199715 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:09 crc kubenswrapper[5035]: E1124 19:15:09.199846 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.253895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.253957 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.253981 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.254012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.254033 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.357113 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.357172 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.357188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.357208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.357226 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.459968 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.460007 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.460018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.460034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.460044 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.563234 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.563268 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.563279 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.563323 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.563337 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.665841 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.665870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.665882 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.665898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.665909 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.767737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.767803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.767837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.767864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.767884 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.870702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.870756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.870775 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.870819 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.870847 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.973439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.973483 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.973494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.973513 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:09 crc kubenswrapper[5035]: I1124 19:15:09.973526 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:09Z","lastTransitionTime":"2025-11-24T19:15:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.076465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.076533 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.076550 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.076575 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.076597 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.184621 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.184715 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.184737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.184765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.184787 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.200102 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:10 crc kubenswrapper[5035]: E1124 19:15:10.200338 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.200743 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:10 crc kubenswrapper[5035]: E1124 19:15:10.200932 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.287676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.287738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.287760 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.287789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.287809 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.390996 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.391058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.391095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.391127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.391143 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.494333 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.494427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.494454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.494487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.494509 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.597377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.597428 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.597444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.597462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.597477 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.700356 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.700433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.700452 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.700477 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.700493 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.803163 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.803235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.803257 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.803341 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.803370 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.906751 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.906802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.906819 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.906843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:10 crc kubenswrapper[5035]: I1124 19:15:10.906861 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:10Z","lastTransitionTime":"2025-11-24T19:15:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.011516 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.011579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.011599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.011624 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.011641 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.114906 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.114998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.115022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.115055 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.115084 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.199942 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.200019 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:11 crc kubenswrapper[5035]: E1124 19:15:11.200150 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:11 crc kubenswrapper[5035]: E1124 19:15:11.200247 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.217943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.218004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.218023 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.218049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.218066 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.321021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.321412 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.321643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.321901 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.322134 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.425197 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.425246 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.425259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.425282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.425319 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.528799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.529106 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.529485 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.529827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.530120 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.633558 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.633622 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.633643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.633677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.633735 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.737045 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.737129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.737151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.737178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.737196 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.840256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.840358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.840381 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.840409 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.840432 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.943373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.943671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.943833 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.943956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:11 crc kubenswrapper[5035]: I1124 19:15:11.944101 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:11Z","lastTransitionTime":"2025-11-24T19:15:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.047270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.047350 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.047369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.047390 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.047405 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.149663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.149860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.149963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.150052 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.150144 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.199832 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.199866 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:12 crc kubenswrapper[5035]: E1124 19:15:12.201030 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.201076 5035 scope.go:117] "RemoveContainer" containerID="9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b" Nov 24 19:15:12 crc kubenswrapper[5035]: E1124 19:15:12.202387 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.225263 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.241634 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.252940 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.252975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.252984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.252998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.253007 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.258447 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.278539 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.295819 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4b6f308065f17325782b9122dc7618140cbcb4ae96d2309d37cdbdec46ec815\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:57Z\\\",\\\"message\\\":\\\"311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.492927 6281 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:14:57.492833 6281 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.493622 6281 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:14:57.494094 6281 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:14:57.494139 6281 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 19:14:57.494150 6281 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 19:14:57.494200 6281 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 19:14:57.494247 6281 factory.go:656] Stopping watch factory\\\\nI1124 19:14:57.494269 6281 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:14:57.494325 6281 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 19:14:57.494346 6281 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 19:14:57.494359 6281 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 19:14:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.310508 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.324387 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.342860 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.356855 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.356904 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.356921 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.356945 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.356961 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.360746 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.378552 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.393512 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.410060 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.426881 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.436761 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.447105 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.460637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.460667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.460677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.460692 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.460702 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.462932 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.480576 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.491714 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.501372 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.511489 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.529006 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.544184 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.556691 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.562579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.562601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.562609 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.562621 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.562630 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.573057 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.589051 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.607868 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.628727 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.643182 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.657190 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.664857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.664967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.665047 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.665114 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.665170 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.674589 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.690081 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.704407 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.716185 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.770759 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.771008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.771169 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.771344 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.771503 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.874576 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.874787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.874933 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.875007 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.875069 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.949813 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/1.log" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.952527 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.953637 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.978988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.979057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.979458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.979478 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.979513 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:12Z","lastTransitionTime":"2025-11-24T19:15:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.979397 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:12 crc kubenswrapper[5035]: I1124 19:15:12.992500 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:12Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.006978 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.026805 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.046687 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.060623 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.072677 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.083459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.083503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.083515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.083531 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.083540 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.086483 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.100990 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.114128 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.124569 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.136370 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.149710 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.159603 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.169404 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.181086 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.185978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.186017 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.186027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.186046 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.186057 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.199459 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.199472 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.199573 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.199662 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.288725 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.288765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.288774 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.288787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.288799 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.391743 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.391792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.391803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.391821 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.391832 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.494394 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.494452 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.494466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.494486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.494498 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.597901 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.597984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.598005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.598026 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.598043 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.701088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.701134 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.701147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.701164 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.701173 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.804145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.804222 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.804242 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.804271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.804324 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.907849 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.907925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.907995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.908025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.908043 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:13Z","lastTransitionTime":"2025-11-24T19:15:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.944377 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.944465 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.944515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.944553 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944693 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944698 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944703 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944763 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:45.94474039 +0000 UTC m=+84.467246687 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944776 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944805 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944724 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944831 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944868 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:45.944856783 +0000 UTC m=+84.467363040 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.944911 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:45.944877434 +0000 UTC m=+84.467383741 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.945017 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.945125 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:45.94506276 +0000 UTC m=+84.467569117 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.960128 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/2.log" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.961160 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/1.log" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.966487 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a" exitCode=1 Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.966636 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a"} Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.966761 5035 scope.go:117] "RemoveContainer" containerID="9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b" Nov 24 19:15:13 crc kubenswrapper[5035]: I1124 19:15:13.967920 5035 scope.go:117] "RemoveContainer" containerID="ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a" Nov 24 19:15:13 crc kubenswrapper[5035]: E1124 19:15:13.968229 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.001022 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:13Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.033617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.033670 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.033691 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.033720 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.033743 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.037877 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.045571 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:15:14 crc kubenswrapper[5035]: E1124 19:15:14.045694 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:15:46.04565359 +0000 UTC m=+84.568159887 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.065263 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.087774 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.103116 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.115400 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.128524 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.136091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.136126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.136162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.136181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.136191 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.142483 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.152643 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.168758 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.183261 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.185041 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.193668 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.199391 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:14 crc kubenswrapper[5035]: E1124 19:15:14.199466 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.199532 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:14 crc kubenswrapper[5035]: E1124 19:15:14.199681 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.207941 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.230406 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.239308 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.239335 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.239346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.239363 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.239374 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.252453 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.271403 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.289191 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.306070 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.336414 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fb94991a430e2c3e9f1400e0ffbaf370845e13cbad5bcd1303aac3c43ee0e0b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"message\\\":\\\".690563 6464 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:14:58Z is after 2025-08-24T17:21:41Z]\\\\nI1124 19:14:58.690775 6464 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver/check-endpoints]} name:Service_openshift-apiserver/check-endpoints_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.139:17698:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.341835 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.341883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.341918 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.341942 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.341958 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.356431 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.373234 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.390535 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.408215 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.426485 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.445226 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.445325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.445380 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.445407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.445425 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.445432 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.465762 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.490390 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.504316 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.516176 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.528764 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.542621 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.547315 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.547357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.547372 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.547393 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.547409 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.556390 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.567680 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.581163 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.649235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.649266 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.649275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.649300 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.649309 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.751190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.751215 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.751223 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.751236 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.751244 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.854204 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.854266 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.854287 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.854350 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.854375 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.957445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.957547 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.957566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.957591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.957607 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:14Z","lastTransitionTime":"2025-11-24T19:15:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.971807 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/2.log" Nov 24 19:15:14 crc kubenswrapper[5035]: I1124 19:15:14.976981 5035 scope.go:117] "RemoveContainer" containerID="ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a" Nov 24 19:15:14 crc kubenswrapper[5035]: E1124 19:15:14.977364 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.002027 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:14Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.019397 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.034904 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.054944 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.060594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.060636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.060648 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.060665 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.060678 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.076603 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.098092 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.113221 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.133446 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.155030 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.162654 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.162707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.162720 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.162738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.162750 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.176552 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.187137 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.199435 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.199435 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:15 crc kubenswrapper[5035]: E1124 19:15:15.199590 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:15 crc kubenswrapper[5035]: E1124 19:15:15.199712 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.200645 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.221139 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.233029 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.245133 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.259017 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.269033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.269092 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.269104 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.269119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.269131 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.273853 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:15Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.371366 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.371398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.371406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.371420 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.371428 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.473906 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.473940 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.473949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.473963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.473972 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.576997 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.577085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.577109 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.577139 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.577162 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.680732 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.680818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.680850 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.680883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.680907 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.784883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.784960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.784987 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.785014 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.785033 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.888264 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.888399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.888422 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.888454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.888476 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.991466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.991514 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.991526 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.991545 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:15 crc kubenswrapper[5035]: I1124 19:15:15.991558 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:15Z","lastTransitionTime":"2025-11-24T19:15:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.093827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.093883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.093897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.093915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.093926 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.196892 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.196934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.196943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.196959 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.196968 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.199329 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.199359 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:16 crc kubenswrapper[5035]: E1124 19:15:16.199573 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:16 crc kubenswrapper[5035]: E1124 19:15:16.199784 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.300875 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.300946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.300969 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.300999 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.301024 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.403779 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.403815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.403825 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.403840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.403850 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.506698 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.506781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.506804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.506835 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.506856 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.609788 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.609838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.609846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.609862 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.609873 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.712635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.712702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.712713 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.712727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.712735 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.818203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.818270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.818331 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.818365 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.818386 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.922006 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.922078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.922096 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.922123 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:16 crc kubenswrapper[5035]: I1124 19:15:16.922140 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:16Z","lastTransitionTime":"2025-11-24T19:15:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.024662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.024716 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.024728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.024748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.024764 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.127746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.127813 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.127832 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.127857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.127890 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.181840 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:17 crc kubenswrapper[5035]: E1124 19:15:17.182008 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:17 crc kubenswrapper[5035]: E1124 19:15:17.182069 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:15:33.182050497 +0000 UTC m=+71.704556764 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.200040 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.200155 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:17 crc kubenswrapper[5035]: E1124 19:15:17.200264 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:17 crc kubenswrapper[5035]: E1124 19:15:17.200381 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.230811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.230854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.230862 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.230877 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.230888 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.333617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.333694 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.333723 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.333752 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.333773 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.437983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.438025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.438036 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.438054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.438065 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.540269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.540339 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.540349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.540363 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.540373 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.643927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.643974 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.643990 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.644018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.644037 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.747719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.747766 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.747778 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.747795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.747807 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.850573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.850633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.850652 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.850682 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.850700 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.954097 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.954152 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.954168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.954189 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:17 crc kubenswrapper[5035]: I1124 19:15:17.954208 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:17Z","lastTransitionTime":"2025-11-24T19:15:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.057477 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.057532 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.057540 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.057556 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.057567 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.160575 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.160610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.160619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.160634 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.160643 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.200051 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.200179 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.200051 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.200291 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.262427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.262473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.262484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.262499 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.262508 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.365005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.365029 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.365037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.365050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.365059 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.467946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.468200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.468312 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.468485 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.468588 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.571229 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.571266 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.571304 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.571319 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.571329 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.670558 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.670607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.670617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.670635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.670645 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.690275 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.695043 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.695170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.695270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.695438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.695544 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.710657 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.715689 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.715754 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.715765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.715781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.715793 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.729112 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.732949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.733012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.733031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.733056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.733075 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.746630 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.751040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.751119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.751170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.751195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.751213 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.768833 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:18Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:18 crc kubenswrapper[5035]: E1124 19:15:18.768951 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.770963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.771008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.771023 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.771042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.771055 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.873722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.873776 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.873788 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.873806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.873817 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.976059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.976094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.976104 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.976117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:18 crc kubenswrapper[5035]: I1124 19:15:18.976127 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:18Z","lastTransitionTime":"2025-11-24T19:15:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.078923 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.079170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.079179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.079195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.079203 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.182193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.182229 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.182238 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.182252 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.182263 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.199808 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.199910 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:19 crc kubenswrapper[5035]: E1124 19:15:19.199939 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:19 crc kubenswrapper[5035]: E1124 19:15:19.200090 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.284758 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.284803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.284816 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.284833 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.284844 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.387603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.387661 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.387678 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.387702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.387721 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.490415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.490463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.490476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.490494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.490507 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.593794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.593861 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.593885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.593913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.593934 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.697066 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.697130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.697153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.697180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.697201 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.800373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.800452 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.800490 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.800521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.800542 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.903320 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.903372 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.903389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.903411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:19 crc kubenswrapper[5035]: I1124 19:15:19.903428 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:19Z","lastTransitionTime":"2025-11-24T19:15:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.006803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.006845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.006855 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.006867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.006876 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.109773 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.109830 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.109847 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.109869 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.109888 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.199685 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:20 crc kubenswrapper[5035]: E1124 19:15:20.199911 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.200434 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:20 crc kubenswrapper[5035]: E1124 19:15:20.200577 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.211639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.211676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.211690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.211707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.211720 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.314483 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.314550 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.314572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.314605 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.314628 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.418218 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.418265 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.418281 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.418334 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.418350 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.521255 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.521360 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.521385 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.521414 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.521435 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.624834 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.624924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.624949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.624980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.625003 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.728809 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.728882 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.728899 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.728926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.728945 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.832206 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.832263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.832275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.832331 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.832343 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.935044 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.935113 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.935130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.935153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:20 crc kubenswrapper[5035]: I1124 19:15:20.935169 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:20Z","lastTransitionTime":"2025-11-24T19:15:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.037839 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.037967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.037980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.037998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.038011 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.140355 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.140503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.140530 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.140553 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.140570 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.199816 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.199820 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:21 crc kubenswrapper[5035]: E1124 19:15:21.199967 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:21 crc kubenswrapper[5035]: E1124 19:15:21.200069 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.242965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.243032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.243050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.243073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.243090 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.345778 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.345870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.345890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.345913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.345930 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.449391 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.449439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.449455 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.449477 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.449495 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.551606 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.551653 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.551668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.551689 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.551704 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.654404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.654454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.654465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.654485 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.654498 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.757643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.757700 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.757718 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.757742 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.757760 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.861885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.861926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.861944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.861965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.861982 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.965438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.965512 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.965557 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.965585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:21 crc kubenswrapper[5035]: I1124 19:15:21.965604 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:21Z","lastTransitionTime":"2025-11-24T19:15:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.068170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.068225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.068242 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.068266 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.068284 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.171523 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.171617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.171635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.171658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.171676 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.199121 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.199173 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:22 crc kubenswrapper[5035]: E1124 19:15:22.199364 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:22 crc kubenswrapper[5035]: E1124 19:15:22.199607 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.217336 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.238987 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.251675 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.263009 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.273283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.273374 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.273424 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.273491 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.273517 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.276361 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.289929 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.310921 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.325730 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.343663 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.360750 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.375059 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.378800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.378860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.378877 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.378900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.378915 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.388806 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.402951 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.414603 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.429878 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.445715 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.460904 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:22Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.481890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.481958 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.481977 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.482442 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.482708 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.585923 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.585989 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.586007 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.586030 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.586056 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.689875 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.690003 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.690025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.690053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.690074 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.792891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.792952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.792970 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.792998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.793017 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.897033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.897455 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.897778 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.898014 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:22 crc kubenswrapper[5035]: I1124 19:15:22.898227 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:22Z","lastTransitionTime":"2025-11-24T19:15:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.000520 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.000614 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.000638 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.000776 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.000798 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.103458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.103518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.103540 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.103566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.103586 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.199525 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:23 crc kubenswrapper[5035]: E1124 19:15:23.199686 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.199971 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:23 crc kubenswrapper[5035]: E1124 19:15:23.200079 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.208397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.208644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.208664 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.208687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.208702 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.311613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.311697 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.311708 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.311726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.311736 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.415086 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.415150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.415167 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.415191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.415208 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.518681 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.518749 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.518769 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.518793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.518811 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.621585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.621644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.621662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.621688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.621706 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.724938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.725010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.725024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.725042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.725081 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.827160 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.827227 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.827247 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.827272 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.827317 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.929605 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.929660 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.929678 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.929697 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:23 crc kubenswrapper[5035]: I1124 19:15:23.929710 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:23Z","lastTransitionTime":"2025-11-24T19:15:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.032806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.032874 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.032888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.032916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.032930 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.135407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.135485 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.135507 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.135530 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.135550 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.200229 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.200349 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:24 crc kubenswrapper[5035]: E1124 19:15:24.200378 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:24 crc kubenswrapper[5035]: E1124 19:15:24.200549 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.237881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.237908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.237917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.237950 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.237961 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.340668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.340722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.340735 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.340752 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.340765 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.442798 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.442849 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.442860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.442876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.442889 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.546313 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.546362 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.546379 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.546399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.546417 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.648986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.649041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.649055 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.649072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.649084 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.751050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.751086 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.751097 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.751113 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.751123 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.853604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.853642 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.853650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.853664 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.853674 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.955991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.956022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.956032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.956047 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:24 crc kubenswrapper[5035]: I1124 19:15:24.956058 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:24Z","lastTransitionTime":"2025-11-24T19:15:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.059544 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.059585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.059599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.059615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.059627 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.161983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.162245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.162374 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.162511 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.162596 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.199810 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:25 crc kubenswrapper[5035]: E1124 19:15:25.200286 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.199810 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:25 crc kubenswrapper[5035]: E1124 19:15:25.201050 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.265417 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.265655 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.265734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.265813 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.265889 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.368938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.369001 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.369026 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.369058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.369080 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.471743 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.471811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.471829 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.471854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.471871 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.575083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.575508 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.575708 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.575904 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.576099 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.678839 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.679062 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.679123 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.679186 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.679250 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.781799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.781867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.781889 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.781917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.781940 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.884888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.884949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.884973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.885004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.885027 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.988199 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.988258 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.988276 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.988326 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:25 crc kubenswrapper[5035]: I1124 19:15:25.988345 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:25Z","lastTransitionTime":"2025-11-24T19:15:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.091089 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.091117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.091125 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.091137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.091145 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.193956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.193994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.194005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.194020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.194031 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.199949 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.199992 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:26 crc kubenswrapper[5035]: E1124 19:15:26.200110 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:26 crc kubenswrapper[5035]: E1124 19:15:26.200271 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.296228 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.296549 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.296561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.296581 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.296593 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.398526 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.398584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.398642 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.398673 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.398693 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.501380 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.501447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.501469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.501498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.501523 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.604977 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.605015 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.605031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.605055 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.605071 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.707426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.707456 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.707467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.707485 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.707499 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.810132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.810175 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.810190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.810212 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.810229 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.912063 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.912099 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.912110 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.912145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:26 crc kubenswrapper[5035]: I1124 19:15:26.912157 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:26Z","lastTransitionTime":"2025-11-24T19:15:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.014768 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.014803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.014822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.014837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.014847 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.116998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.117058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.117068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.117081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.117091 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.199496 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:27 crc kubenswrapper[5035]: E1124 19:15:27.199631 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.199497 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:27 crc kubenswrapper[5035]: E1124 19:15:27.199798 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.220022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.220061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.220078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.220099 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.220114 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.322866 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.322923 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.322934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.322951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.322963 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.425126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.425166 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.425179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.425198 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.425210 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.527871 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.527910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.527921 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.527962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.527975 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.630433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.630466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.630477 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.630493 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.630506 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.732233 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.732265 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.732275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.732305 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.732317 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.834787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.834826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.834837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.834853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.834864 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.938458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.938522 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.938539 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.938556 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:27 crc kubenswrapper[5035]: I1124 19:15:27.938568 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:27Z","lastTransitionTime":"2025-11-24T19:15:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.040848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.040886 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.040898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.040915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.040928 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.142916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.142964 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.142978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.142993 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.143004 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.199706 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.199790 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.199937 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.200013 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.200551 5035 scope.go:117] "RemoveContainer" containerID="ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a" Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.200764 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.244644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.244668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.244678 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.244691 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.244703 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.347308 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.347361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.347376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.347395 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.347405 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.449659 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.449696 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.449705 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.449718 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.449746 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.553034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.553069 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.553081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.553095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.553106 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.655053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.655089 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.655100 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.655114 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.655126 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.756624 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.756901 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.756990 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.757077 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.757160 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.850467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.850611 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.850693 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.850770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.850867 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.870711 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.875107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.875241 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.875358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.875448 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.875540 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.892567 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.896667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.896804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.896988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.897082 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.897165 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.917655 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.921459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.921583 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.921671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.921757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.921840 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.934891 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.937865 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.937902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.937913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.937928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.937938 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.952483 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:28Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:28 crc kubenswrapper[5035]: E1124 19:15:28.952591 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.954317 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.954466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.954553 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.954648 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:28 crc kubenswrapper[5035]: I1124 19:15:28.954727 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:28Z","lastTransitionTime":"2025-11-24T19:15:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.056190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.056576 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.056683 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.056765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.056845 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.159966 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.160016 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.160034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.160056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.160072 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.199449 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:29 crc kubenswrapper[5035]: E1124 19:15:29.200028 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.199443 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:29 crc kubenswrapper[5035]: E1124 19:15:29.203427 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.262388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.262439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.262456 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.262480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.262498 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.365974 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.366019 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.366035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.366057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.366073 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.471061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.471116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.471133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.471155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.471173 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.573402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.573465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.573483 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.573509 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.573527 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.676548 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.676644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.676680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.676720 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.676742 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.779811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.779869 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.779887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.779910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.779932 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.882649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.882722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.882738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.882764 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.882780 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.985837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.985939 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.985956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.985979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:29 crc kubenswrapper[5035]: I1124 19:15:29.985995 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:29Z","lastTransitionTime":"2025-11-24T19:15:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.088406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.088467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.088484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.088505 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.088522 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.191078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.191116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.191127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.191142 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.191153 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.199483 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.199531 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:30 crc kubenswrapper[5035]: E1124 19:15:30.199671 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:30 crc kubenswrapper[5035]: E1124 19:15:30.199806 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.293891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.294206 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.294314 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.294427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.294513 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.397611 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.397649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.397661 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.397675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.397686 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.499854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.499929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.499955 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.499984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.500003 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.602188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.602254 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.602269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.602373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.602388 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.704273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.704357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.704373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.704395 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.704407 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.806523 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.806927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.807191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.807457 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.807711 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.911016 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.911049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.911058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.911072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:30 crc kubenswrapper[5035]: I1124 19:15:30.911080 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:30Z","lastTransitionTime":"2025-11-24T19:15:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.014067 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.014115 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.014127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.014143 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.014153 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.116776 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.116827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.116838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.116854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.116864 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.200010 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.200023 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:31 crc kubenswrapper[5035]: E1124 19:15:31.200235 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:31 crc kubenswrapper[5035]: E1124 19:15:31.200420 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.218831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.218864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.218876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.218890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.218902 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.321680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.321746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.321769 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.321798 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.321821 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.424189 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.424242 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.424259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.424282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.424325 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.526302 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.526334 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.526345 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.526358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.526367 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.628956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.629015 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.629037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.629064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.629088 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.731783 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.731853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.731870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.731894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.731921 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.833834 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.833878 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.833899 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.833928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.833948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.936138 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.936200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.936211 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.936225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:31 crc kubenswrapper[5035]: I1124 19:15:31.936238 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:31Z","lastTransitionTime":"2025-11-24T19:15:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.042511 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.042584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.042601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.042625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.042645 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.145177 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.145225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.145240 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.145259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.145273 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.200121 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:32 crc kubenswrapper[5035]: E1124 19:15:32.200434 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.200736 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:32 crc kubenswrapper[5035]: E1124 19:15:32.200836 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.218067 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.232895 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.248772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.248816 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.248831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.248853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.248828 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.248871 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.265997 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.283811 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.302938 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.322869 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.338813 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.351935 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.351967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.351978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.351992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.352038 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.353892 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.369368 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.388006 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.403050 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.415456 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.427708 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.442881 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.454102 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.454127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.454135 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.454147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.454156 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.456283 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.474518 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:32Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.555988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.556036 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.556049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.556084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.556097 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.659165 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.659531 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.659591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.659629 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.659657 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.762404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.762446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.762454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.762470 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.762479 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.864991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.865031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.865040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.865055 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.865063 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.967719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.967791 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.967803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.967826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:32 crc kubenswrapper[5035]: I1124 19:15:32.967841 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:32Z","lastTransitionTime":"2025-11-24T19:15:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.069806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.069876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.069901 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.069929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.069951 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.173325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.173361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.173379 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.173397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.173411 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.199649 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.199662 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:33 crc kubenswrapper[5035]: E1124 19:15:33.199863 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:33 crc kubenswrapper[5035]: E1124 19:15:33.199998 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.209098 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.253760 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:33 crc kubenswrapper[5035]: E1124 19:15:33.253923 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:33 crc kubenswrapper[5035]: E1124 19:15:33.253983 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:05.253964988 +0000 UTC m=+103.776471245 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.276021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.276060 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.276071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.276087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.276100 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.378674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.378728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.378746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.378770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.378790 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.481612 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.481650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.481659 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.481674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.481684 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.583396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.583434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.583448 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.583485 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.583498 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.685583 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.685647 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.685658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.685677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.685709 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.788354 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.788394 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.788403 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.788418 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.788428 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.890361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.890402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.890413 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.890425 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.890434 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.992895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.992932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.992943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.992962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:33 crc kubenswrapper[5035]: I1124 19:15:33.992974 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:33Z","lastTransitionTime":"2025-11-24T19:15:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.095721 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.095759 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.095770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.095787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.095801 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.198433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.198484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.198499 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.198522 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.198543 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.199096 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.199185 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:34 crc kubenswrapper[5035]: E1124 19:15:34.199237 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:34 crc kubenswrapper[5035]: E1124 19:15:34.199490 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.302011 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.302061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.302077 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.302099 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.302116 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.404853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.404896 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.404909 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.404927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.404939 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.507826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.507895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.507923 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.507984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.508008 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.611506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.611582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.611606 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.611632 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.611649 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.714354 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.714426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.714450 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.714481 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.714502 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.817534 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.817594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.817619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.817652 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.817674 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.920810 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.920878 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.920900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.920927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:34 crc kubenswrapper[5035]: I1124 19:15:34.920948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:34Z","lastTransitionTime":"2025-11-24T19:15:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.022747 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.022794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.022811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.022835 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.022858 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.125479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.125544 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.125561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.125587 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.125606 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.199451 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.199502 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:35 crc kubenswrapper[5035]: E1124 19:15:35.199597 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:35 crc kubenswrapper[5035]: E1124 19:15:35.199724 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.227919 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.228002 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.228021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.228053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.228072 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.331196 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.331241 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.331253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.331271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.331283 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.434133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.434215 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.434242 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.434267 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.434285 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.537770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.537827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.537839 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.537856 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.537871 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.640482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.641358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.641377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.641400 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.641409 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.743773 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.743819 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.743832 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.743848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.743859 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.846945 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.847008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.847047 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.847073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.847089 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.950406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.950469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.950486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.950510 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:35 crc kubenswrapper[5035]: I1124 19:15:35.950526 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:35Z","lastTransitionTime":"2025-11-24T19:15:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.052707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.052778 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.052790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.052831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.052844 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.156986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.157137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.157153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.157173 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.157187 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.199928 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:36 crc kubenswrapper[5035]: E1124 19:15:36.200173 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.200455 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:36 crc kubenswrapper[5035]: E1124 19:15:36.200556 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.261096 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.261141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.261158 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.261182 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.261199 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.364559 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.364597 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.364607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.364622 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.364633 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.467517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.467561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.467578 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.467600 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.467616 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.570794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.570840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.570856 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.570879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.570896 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.673577 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.673637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.673662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.673692 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.673713 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.777140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.777194 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.777210 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.777233 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.777250 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.879317 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.879373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.879386 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.879402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.879413 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.982896 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.982964 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.982989 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.983020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:36 crc kubenswrapper[5035]: I1124 19:15:36.983048 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:36Z","lastTransitionTime":"2025-11-24T19:15:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.056770 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/0.log" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.056824 5035 generic.go:334] "Generic (PLEG): container finished" podID="38b324c3-dc7d-4555-ac0d-714a9f1a40cf" containerID="6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c" exitCode=1 Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.056856 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerDied","Data":"6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.057252 5035 scope.go:117] "RemoveContainer" containerID="6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.072230 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.085507 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.099134 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.103889 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.103926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.103942 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.103959 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.103973 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.112384 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.126022 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.138390 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.148958 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.158604 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.170561 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.183360 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.195093 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.206778 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.207088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.207117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.207133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.207157 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.207173 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.219282 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.232651 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.244444 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.417938 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:37 crc kubenswrapper[5035]: E1124 19:15:37.418088 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.418333 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:37 crc kubenswrapper[5035]: E1124 19:15:37.418417 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.423026 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.423094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.423116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.423144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.423172 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.444444 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.464824 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.479799 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:37Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.527848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.527908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.527926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.528012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.528090 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.631927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.632010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.632022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.632046 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.632060 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.734857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.734914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.734934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.734958 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.734976 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.838721 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.838794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.838808 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.838833 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.838849 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.941683 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.941791 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.941815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.941846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:37 crc kubenswrapper[5035]: I1124 19:15:37.941869 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:37Z","lastTransitionTime":"2025-11-24T19:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.045075 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.045131 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.045155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.045453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.045490 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.062185 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/0.log" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.062248 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerStarted","Data":"91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.094512 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.109448 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.123352 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.135822 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.147859 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.147892 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.147902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.147915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.147926 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.151167 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.165492 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.177990 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.193784 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.199962 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:38 crc kubenswrapper[5035]: E1124 19:15:38.200075 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.200244 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:38 crc kubenswrapper[5035]: E1124 19:15:38.200342 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.214420 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.225642 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.235818 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.250119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.250160 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.250173 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.250095 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.250188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.250331 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.266972 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.285920 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.300424 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.318960 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.334572 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.346438 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:38Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.352995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.353032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.353041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.353055 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.353065 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.456424 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.456476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.456491 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.456514 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.456530 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.559617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.559678 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.559696 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.559720 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.559738 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.662754 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.662813 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.662830 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.662854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.662874 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.766451 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.766509 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.766526 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.766551 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.766568 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.869424 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.869480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.869497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.869521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.869539 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.973001 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.973064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.973083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.973109 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:38 crc kubenswrapper[5035]: I1124 19:15:38.973126 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:38Z","lastTransitionTime":"2025-11-24T19:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.075914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.075977 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.075995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.076022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.076039 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.179173 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.179231 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.179248 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.179272 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.179329 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.200115 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.200266 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.200386 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.200445 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.281727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.281789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.281807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.281836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.281856 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.350054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.350100 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.350117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.350136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.350152 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.372188 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.377382 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.377430 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.377446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.377467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.377483 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.400237 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.404599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.404649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.404668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.404690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.404706 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.422624 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.427355 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.427417 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.427441 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.427470 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.427492 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.449719 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.454972 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.455031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.455056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.455086 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.455108 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.472400 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:39Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:39 crc kubenswrapper[5035]: E1124 19:15:39.472622 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.474684 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.474737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.474756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.474783 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.474805 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.576562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.576597 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.576605 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.576618 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.576627 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.679961 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.680010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.680024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.680042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.680056 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.782046 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.782072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.782079 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.782092 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.782100 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.884763 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.884831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.884845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.884865 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.884880 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.988124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.988175 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.988189 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.988209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:39 crc kubenswrapper[5035]: I1124 19:15:39.988223 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:39Z","lastTransitionTime":"2025-11-24T19:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.090473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.090551 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.090566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.090590 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.090606 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.193497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.193559 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.193583 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.193614 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.193638 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.199162 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:40 crc kubenswrapper[5035]: E1124 19:15:40.199403 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.199179 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:40 crc kubenswrapper[5035]: E1124 19:15:40.199758 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.296600 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.296909 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.297091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.297353 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.297889 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.402053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.402126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.402144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.402170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.402189 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.505075 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.505444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.505607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.505932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.506092 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.609477 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.609553 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.609574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.609602 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.609621 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.713629 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.714131 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.714385 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.714604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.714919 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.818377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.818456 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.818475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.819447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.819541 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.923717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.924116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.924360 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.924560 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:40 crc kubenswrapper[5035]: I1124 19:15:40.924733 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:40Z","lastTransitionTime":"2025-11-24T19:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.029678 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.029741 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.029765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.029798 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.029819 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.132644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.132718 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.132734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.132758 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.132776 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.199504 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.199569 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:41 crc kubenswrapper[5035]: E1124 19:15:41.199705 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:41 crc kubenswrapper[5035]: E1124 19:15:41.199853 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.235548 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.235604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.235620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.235643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.235665 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.339552 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.339619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.339637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.339668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.339690 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.442429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.442489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.442510 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.442533 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.442551 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.545862 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.545926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.545949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.545979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.546001 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.649389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.649460 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.649479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.649506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.649526 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.752397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.752742 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.752915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.753087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.753235 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.856921 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.858503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.858741 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.858924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.859108 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.961248 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.961273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.961281 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.961316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:41 crc kubenswrapper[5035]: I1124 19:15:41.961326 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:41Z","lastTransitionTime":"2025-11-24T19:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.064594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.064667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.064690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.064718 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.064738 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.168633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.169012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.169169 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.169383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.169579 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.199158 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.199281 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:42 crc kubenswrapper[5035]: E1124 19:15:42.199377 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:42 crc kubenswrapper[5035]: E1124 19:15:42.199502 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.200258 5035 scope.go:117] "RemoveContainer" containerID="ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.227535 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.257117 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.277118 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.277197 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.277159 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.277225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.277415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.277436 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.297218 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.317624 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.337586 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.356422 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.373754 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.379807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.379957 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.380061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.380179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.380266 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.389917 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.404333 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.423245 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.440429 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.459447 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.475760 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.482487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.482508 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.482516 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.482529 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.482537 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.494897 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.507330 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.522127 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.551821 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:42Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.585072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.585134 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.585151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.585175 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.585192 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.688646 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.688699 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.688724 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.688756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.688775 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.791628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.791658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.791671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.791686 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.791697 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.893556 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.893582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.893589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.893603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.893611 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.996367 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.996425 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.996435 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.996449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:42 crc kubenswrapper[5035]: I1124 19:15:42.996458 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:42Z","lastTransitionTime":"2025-11-24T19:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.082072 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/2.log" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.084702 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.085195 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.098275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.098482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.098570 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.098632 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.098693 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.105449 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.117219 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.128368 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.145510 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.156712 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.171959 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.182683 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.199093 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.199361 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:43 crc kubenswrapper[5035]: E1124 19:15:43.199354 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:43 crc kubenswrapper[5035]: E1124 19:15:43.199633 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.200370 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.200400 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.200410 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.200425 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.200436 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.202868 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.217637 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.232994 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.250427 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.261075 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.272063 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.284912 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.299081 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.303158 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.303208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.303224 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.303245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.303260 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.311585 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.319977 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.330315 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:43Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.406433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.406466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.406476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.406494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.406506 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.509910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.510227 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.510447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.510650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.510788 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.613592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.613638 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.613652 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.613674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.613689 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.716206 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.716247 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.716259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.716274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.716284 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.819997 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.820063 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.820084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.820107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.820124 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.923368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.923411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.923422 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.923439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:43 crc kubenswrapper[5035]: I1124 19:15:43.923452 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:43Z","lastTransitionTime":"2025-11-24T19:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.026186 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.026257 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.026275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.026339 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.026366 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.091653 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/3.log" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.092640 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/2.log" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.096959 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d" exitCode=1 Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.097027 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.097086 5035 scope.go:117] "RemoveContainer" containerID="ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.098369 5035 scope.go:117] "RemoveContainer" containerID="6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d" Nov 24 19:15:44 crc kubenswrapper[5035]: E1124 19:15:44.098696 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.119186 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.128860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.128932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.128955 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.128987 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.129008 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.134537 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.148931 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.170505 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae331de983ae2500a9a06e6dbe2b8c08ad265482f607199819b318ff9255404a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:13Z\\\",\\\"message\\\":\\\"pping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098834 6683 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.098873 6683 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098917 6683 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.098995 6683 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 19:15:13.099200 6683 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 19:15:13.099422 6683 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 19:15:13.099455 6683 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 19:15:13.099476 6683 factory.go:656] Stopping watch factory\\\\nI1124 19:15:13.099492 6683 ovnkube.go:599] Stopped ovnkube\\\\nI1124 19:15:13.099513 6683 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 19:15:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:43Z\\\",\\\"message\\\":\\\" [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227125 7098 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 19:15:43.227139 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227162 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 19:15:43.227206 7098 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.190255 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.199246 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:44 crc kubenswrapper[5035]: E1124 19:15:44.199398 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.199524 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:44 crc kubenswrapper[5035]: E1124 19:15:44.200772 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.213247 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.228652 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.231950 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.231971 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.231981 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.231994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.232004 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.246391 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.262367 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.275096 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.288142 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.303940 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.320549 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.334934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.335009 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.335027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.335056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.335073 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.336861 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.351554 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.364282 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.376302 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.388421 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:44Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.439827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.439877 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.439893 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.439914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.439928 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.543675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.543736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.543757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.543781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.543800 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.646992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.647062 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.647078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.647096 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.647107 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.749902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.749972 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.749994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.750021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.750041 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.853497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.853567 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.853593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.853623 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.853645 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.956986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.957036 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.957053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.957077 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:44 crc kubenswrapper[5035]: I1124 19:15:44.957094 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:44Z","lastTransitionTime":"2025-11-24T19:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.059933 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.059991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.060013 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.060042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.060065 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.103803 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/3.log" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.108042 5035 scope.go:117] "RemoveContainer" containerID="6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d" Nov 24 19:15:45 crc kubenswrapper[5035]: E1124 19:15:45.108340 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.122106 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.137578 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.176615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.176694 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.176722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.176756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.176779 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.178446 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.199806 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:45 crc kubenswrapper[5035]: E1124 19:15:45.199962 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.199812 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:45 crc kubenswrapper[5035]: E1124 19:15:45.200238 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.209197 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.221283 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.231942 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.243114 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.254654 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.264320 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.273529 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.279828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.279870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.279881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.279898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.279909 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.285563 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.296029 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.306095 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.324766 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:43Z\\\",\\\"message\\\":\\\" [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227125 7098 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 19:15:43.227139 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227162 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 19:15:43.227206 7098 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.336584 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.348200 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.370156 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.382501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.382644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.382721 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.382756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.382844 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.385803 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:45Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.485368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.485407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.485415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.485429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.485440 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.587757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.587805 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.587823 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.587843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.587856 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.691518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.691588 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.691643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.691675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.691696 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.794883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.794928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.794941 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.794959 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.794973 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.898054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.898108 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.898119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.898136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:45 crc kubenswrapper[5035]: I1124 19:15:45.898148 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:45Z","lastTransitionTime":"2025-11-24T19:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.001266 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.001336 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.001349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.001366 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.001376 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.022971 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.023033 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.023075 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.023103 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023116 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023205 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:50.023184597 +0000 UTC m=+148.545690854 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023275 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023352 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023386 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023390 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023442 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023464 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023405 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023419 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:50.023400372 +0000 UTC m=+148.545906629 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023612 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:50.023595367 +0000 UTC m=+148.546101694 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.023630 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:50.023621108 +0000 UTC m=+148.546127445 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.103932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.103971 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.103979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.104010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.104020 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.124439 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.124604 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:50.124582337 +0000 UTC m=+148.647088594 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.199996 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.200040 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.200167 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:46 crc kubenswrapper[5035]: E1124 19:15:46.200380 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.206898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.206937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.206946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.206963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.206972 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.309557 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.309596 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.309610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.309626 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.309638 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.412053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.412120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.412137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.412161 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.412177 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.514765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.514842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.514865 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.514894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.514918 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.618200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.618268 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.618325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.618363 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.618382 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.721102 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.721227 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.721245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.721275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.721354 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.824331 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.824399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.824425 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.824456 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.824481 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.927835 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.927911 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.927929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.927955 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:46 crc kubenswrapper[5035]: I1124 19:15:46.927972 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:46Z","lastTransitionTime":"2025-11-24T19:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.030708 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.030766 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.030778 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.030796 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.030809 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.134595 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.134673 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.134699 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.134729 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.134754 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.200026 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.200083 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:47 crc kubenswrapper[5035]: E1124 19:15:47.200222 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:47 crc kubenswrapper[5035]: E1124 19:15:47.200409 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.237821 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.237873 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.237886 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.237931 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.237944 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.340768 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.340854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.340871 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.340896 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.340912 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.443606 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.443677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.443719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.443743 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.443762 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.553914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.553973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.553991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.554016 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.554034 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.657404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.657447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.657463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.657487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.657504 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.759673 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.759703 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.759711 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.759726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.759735 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.862751 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.862813 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.862838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.862867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.862887 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.966562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.966605 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.966620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.966640 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:47 crc kubenswrapper[5035]: I1124 19:15:47.966656 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:47Z","lastTransitionTime":"2025-11-24T19:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.069800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.069861 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.069879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.069905 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.069925 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.172634 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.173085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.173389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.173589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.173760 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.200516 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.200578 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:48 crc kubenswrapper[5035]: E1124 19:15:48.200672 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:48 crc kubenswrapper[5035]: E1124 19:15:48.200822 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.276689 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.277150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.277226 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.277331 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.277424 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.380454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.380524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.380543 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.380569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.380586 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.483137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.483635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.486956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.487086 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.487200 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.591358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.591409 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.591426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.591446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.591462 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.693680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.693946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.694147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.694219 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.694320 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.797698 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.797792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.797814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.797843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.797859 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.901064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.901449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.901639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.901807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:48 crc kubenswrapper[5035]: I1124 19:15:48.901981 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:48Z","lastTransitionTime":"2025-11-24T19:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.006134 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.006201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.006221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.006245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.006265 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.109517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.109576 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.109600 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.109630 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.109652 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.199538 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.199727 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.199538 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.200059 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.213473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.213540 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.213561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.213588 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.213611 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.316845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.316895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.316911 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.316937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.316954 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.420712 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.420781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.420803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.420828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.420848 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.524162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.524515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.524648 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.524770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.524888 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.627975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.628038 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.628057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.628095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.628117 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.730584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.730626 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.730637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.730652 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.730662 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.771789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.771835 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.771845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.771859 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.771869 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.791063 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.796411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.796461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.796474 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.796493 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.796506 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.808430 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.819012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.819059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.819071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.819091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.819104 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.838391 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.843525 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.843791 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.844132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.844618 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.844804 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.859260 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.863604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.863641 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.863651 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.863665 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.863675 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.882759 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:49 crc kubenswrapper[5035]: E1124 19:15:49.883438 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.885477 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.885676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.886035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.886230 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.886460 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.988522 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.988866 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.989126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.989431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:49 crc kubenswrapper[5035]: I1124 19:15:49.989620 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:49Z","lastTransitionTime":"2025-11-24T19:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.092915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.092997 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.093022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.093088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.093116 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.195880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.195924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.195937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.195956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.195968 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.199365 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.199409 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:50 crc kubenswrapper[5035]: E1124 19:15:50.199588 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:50 crc kubenswrapper[5035]: E1124 19:15:50.199779 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.298836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.298881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.298916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.298933 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.298944 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.401386 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.401432 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.401443 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.401463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.401478 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.504068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.504177 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.504192 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.504210 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.504222 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.606435 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.606469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.606479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.606494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.606504 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.709254 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.709338 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.709355 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.709376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.709390 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.811613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.811852 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.811954 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.812059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.812385 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.915341 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.915382 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.915397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.915416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:50 crc kubenswrapper[5035]: I1124 19:15:50.915430 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:50Z","lastTransitionTime":"2025-11-24T19:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.017991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.018064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.018083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.018109 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.018124 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.120074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.120121 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.120136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.120156 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.120171 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.199386 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.199455 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:51 crc kubenswrapper[5035]: E1124 19:15:51.200358 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:51 crc kubenswrapper[5035]: E1124 19:15:51.200535 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.222137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.222186 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.222198 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.222217 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.222229 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.325900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.326018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.326037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.326085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.326126 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.429806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.430198 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.430408 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.430574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.430747 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.534573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.534622 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.534637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.534658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.534672 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.638081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.638137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.638159 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.638187 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.638208 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.740661 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.740729 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.740753 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.740785 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.740808 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.843820 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.843859 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.843870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.843888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.843901 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.947024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.947062 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.947072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.947085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:51 crc kubenswrapper[5035]: I1124 19:15:51.947095 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:51Z","lastTransitionTime":"2025-11-24T19:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.049398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.049434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.049444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.049460 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.049472 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.152103 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.152141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.152157 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.152176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.152191 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.199242 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:52 crc kubenswrapper[5035]: E1124 19:15:52.199495 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.200243 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:52 crc kubenswrapper[5035]: E1124 19:15:52.201153 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.224320 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.244185 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.254242 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.254487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.254591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.254698 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.254802 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.263776 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.278520 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.291676 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.305440 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.320051 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.336504 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.348946 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.357610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.357660 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.357672 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.357690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.357704 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.363029 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.377403 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.391972 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.403308 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.417188 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.432086 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.460881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.460929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.460953 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.460981 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.461005 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.471531 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.486373 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.514536 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:43Z\\\",\\\"message\\\":\\\" [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227125 7098 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 19:15:43.227139 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227162 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 19:15:43.227206 7098 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.563864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.563914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.563932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.563955 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.563974 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.668156 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.668259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.668276 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.668316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.668337 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.772123 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.772891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.773121 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.773273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.773468 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.876921 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.877260 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.877665 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.878147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.878515 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.982188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.982599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.982739 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.982876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:52 crc kubenswrapper[5035]: I1124 19:15:52.982969 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:52Z","lastTransitionTime":"2025-11-24T19:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.085029 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.085263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.085387 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.085466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.085525 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.187568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.187810 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.187881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.187951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.188012 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.199839 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.199839 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:53 crc kubenswrapper[5035]: E1124 19:15:53.200130 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:53 crc kubenswrapper[5035]: E1124 19:15:53.200053 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.290221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.290754 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.290831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.290894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.290948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.394271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.394402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.394423 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.394448 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.394468 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.498075 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.498218 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.498237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.498270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.498287 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.601220 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.601522 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.601550 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.601579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.601601 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.704978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.705056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.705081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.705112 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.705137 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.809194 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.809267 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.809286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.809362 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.809380 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.912741 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.912790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.912803 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.912823 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:53 crc kubenswrapper[5035]: I1124 19:15:53.912837 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:53Z","lastTransitionTime":"2025-11-24T19:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.016407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.016455 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.016468 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.016491 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.016505 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.120090 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.120130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.120150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.120176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.120194 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.200024 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.200044 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:54 crc kubenswrapper[5035]: E1124 19:15:54.200227 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:54 crc kubenswrapper[5035]: E1124 19:15:54.200325 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.223024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.223064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.223077 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.223095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.223107 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.325556 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.325592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.325603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.325621 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.325637 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.428572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.428670 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.428688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.428749 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.428769 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.531867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.532140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.532233 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.532348 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.532430 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.634894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.634927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.634937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.634952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.634963 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.737807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.737882 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.737900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.737930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.737948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.840860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.840915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.840933 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.840957 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.840975 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.944660 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.944736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.944756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.944782 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:54 crc kubenswrapper[5035]: I1124 19:15:54.944802 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:54Z","lastTransitionTime":"2025-11-24T19:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.047719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.047786 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.047810 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.047836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.047853 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.150483 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.150521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.150535 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.150551 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.150561 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.199367 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.199466 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:55 crc kubenswrapper[5035]: E1124 19:15:55.199972 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:55 crc kubenswrapper[5035]: E1124 19:15:55.200007 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.252994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.253040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.253054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.253073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.253085 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.356020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.356070 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.356085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.356108 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.356121 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.458915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.458980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.459001 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.459027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.459085 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.562140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.562202 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.562224 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.562254 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.562276 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.666582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.666973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.667270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.667528 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.667782 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.774838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.774917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.774945 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.774976 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.774998 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.878079 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.878415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.878625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.878795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.878946 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.982465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.982887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.983058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.983221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:55 crc kubenswrapper[5035]: I1124 19:15:55.983416 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:55Z","lastTransitionTime":"2025-11-24T19:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.086948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.086985 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.086995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.087011 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.087022 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.189726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.189777 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.189788 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.189802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.189833 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.199355 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.199468 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:56 crc kubenswrapper[5035]: E1124 19:15:56.199605 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:56 crc kubenswrapper[5035]: E1124 19:15:56.199726 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.291954 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.292011 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.292027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.292051 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.292068 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.394470 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.395343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.395565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.395784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.395956 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.499450 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.499833 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.499986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.500134 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.500277 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.603205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.603263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.603282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.603340 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.603363 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.706140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.706257 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.706284 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.706346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.706369 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.808767 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.809062 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.809175 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.809369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.809493 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.912717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.912770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.912792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.912821 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:56 crc kubenswrapper[5035]: I1124 19:15:56.912842 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:56Z","lastTransitionTime":"2025-11-24T19:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.015762 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.015795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.015845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.015862 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.015872 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.118832 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.118959 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.118992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.119021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.119042 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.199985 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.200022 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:57 crc kubenswrapper[5035]: E1124 19:15:57.200705 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:57 crc kubenswrapper[5035]: E1124 19:15:57.200573 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.228026 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.228339 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.228706 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.229060 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.229391 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.331921 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.332369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.332568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.332722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.332850 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.435436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.436009 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.436078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.436156 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.436221 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.537978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.538021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.538037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.538061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.538077 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.640220 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.640256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.640268 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.640283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.640315 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.743229 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.743566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.743673 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.743760 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.743835 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.846885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.846931 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.846943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.846961 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.846972 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.949891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.949949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.949963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.949982 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:57 crc kubenswrapper[5035]: I1124 19:15:57.949994 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:57Z","lastTransitionTime":"2025-11-24T19:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.053584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.053883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.053980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.054073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.054134 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.155785 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.155814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.155822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.155838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.155847 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.199788 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.199789 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:15:58 crc kubenswrapper[5035]: E1124 19:15:58.200106 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:15:58 crc kubenswrapper[5035]: E1124 19:15:58.200257 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.258991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.259372 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.259625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.259975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.260259 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.363134 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.363225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.363248 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.363277 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.363338 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.468504 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.468563 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.468581 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.468604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.468621 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.572085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.572408 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.572487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.572556 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.572619 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.676340 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.676693 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.676902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.677171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.677402 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.780687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.780763 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.780799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.780831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.780852 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.883541 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.883600 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.883617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.883641 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.883658 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.986566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.986604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.986615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.986631 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:58 crc kubenswrapper[5035]: I1124 19:15:58.986642 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:58Z","lastTransitionTime":"2025-11-24T19:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.088828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.088868 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.088880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.088897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.088909 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.191402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.191460 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.191479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.191503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.191520 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.199744 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.199788 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:15:59 crc kubenswrapper[5035]: E1124 19:15:59.200190 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:15:59 crc kubenswrapper[5035]: E1124 19:15:59.203651 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.294653 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.294700 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.294725 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.294749 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.294766 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.397536 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.397608 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.397633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.397663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.397688 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.500405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.500945 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.501014 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.501089 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.501156 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.604318 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.604361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.604372 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.604393 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.604407 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.707080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.707144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.707162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.707192 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.707210 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.810273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.810374 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.810404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.810444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.810468 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.913368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.913426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.913442 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.913465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:15:59 crc kubenswrapper[5035]: I1124 19:15:59.913482 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:15:59Z","lastTransitionTime":"2025-11-24T19:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.017320 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.017396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.017414 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.017437 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.017451 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.121067 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.121127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.121148 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.121174 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.121192 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.200064 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.200277 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.200962 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.201263 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.201623 5035 scope.go:117] "RemoveContainer" containerID="6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d" Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.201895 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.211107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.211184 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.211206 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.211239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.211259 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.232657 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.238952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.239025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.239049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.239079 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.239100 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.261366 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.266825 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.266879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.266896 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.266919 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.266939 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.283649 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.287845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.287981 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.288045 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.288108 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.288170 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.304594 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.308766 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.308815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.308836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.308858 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.308875 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.323173 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T19:16:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"397c8539-193a-45fe-895a-942e12dca119\\\",\\\"systemUUID\\\":\\\"17ceffe9-e870-43f4-80f9-b6f15e4ae84e\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:00 crc kubenswrapper[5035]: E1124 19:16:00.327506 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.329706 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.329738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.329749 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.329767 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.329780 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.432287 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.432575 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.432655 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.432757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.432848 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.535916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.535985 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.536008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.536036 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.536058 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.638715 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.638772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.638796 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.638823 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.638847 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.742149 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.742239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.742256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.742379 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.742418 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.845625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.845688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.845706 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.845732 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.845765 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.948734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.948779 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.948797 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.948820 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:00 crc kubenswrapper[5035]: I1124 19:16:00.948838 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:00Z","lastTransitionTime":"2025-11-24T19:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.051530 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.051604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.051628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.051658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.051679 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.154758 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.154802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.154818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.154839 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.154856 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.199049 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.199120 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:01 crc kubenswrapper[5035]: E1124 19:16:01.199205 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:01 crc kubenswrapper[5035]: E1124 19:16:01.199790 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.220380 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.258103 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.258529 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.258703 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.258854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.259039 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.362764 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.362814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.362834 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.362860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.362878 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.465222 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.465274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.465329 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.465362 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.465381 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.568888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.569800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.569989 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.570172 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.570396 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.673321 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.673375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.673395 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.673420 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.673437 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.775874 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.775932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.775951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.775975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.775991 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.879330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.879389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.879406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.879430 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.879447 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.982643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.982938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.983127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.983271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:01 crc kubenswrapper[5035]: I1124 19:16:01.983479 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:01Z","lastTransitionTime":"2025-11-24T19:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.086796 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.086853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.086870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.086893 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.086910 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.188971 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.189021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.189035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.189054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.189066 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.199206 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.199225 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:02 crc kubenswrapper[5035]: E1124 19:16:02.199506 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:02 crc kubenswrapper[5035]: E1124 19:16:02.199642 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.217501 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a8ce0a7-1672-47d7-8b17-9443072b3ef4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://748974afa020db20324dcc7c20b5ae45a35a6650281835f0faa77d7a5628a20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f6edd967269826098ece1dbc3b34890acb27014c5e83654681b7164b2122de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f4130878b881da3d45e13a07e2932be10b4645d47062fe66b861ac50dded58\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eba04d00932a6aabcb30f6a5ffe4de27374b977b1ff1fa595a52f97d9a41b59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.230993 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc3b0830-7575-463a-acba-80b6ce711ee3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0074614ad33b530c0fe18b5bd526480d979aa109355aa954a748ff0408b5e27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4baa0c17477f94b024eb34e40876268643c92d3b1e30d2d8659c9a7912603dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://32bcd1497bbc880cff5c1878a898ab8e43f561d63f586dd4574390daea3604a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b29f5c2d0391f59e84fa9e0f3c4ad0f043b4ed86792136451d58cf61247c0c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.243006 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21abd6f7-fd62-498a-b347-7e74297df351\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e60ed437730054c27a781787f7c20600514f8e522f18eaddc22c458a37b47f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71407e70fb354124222112e4829353d701c867a9ce000f429cee98aae5bc0105\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.263775 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:43Z\\\",\\\"message\\\":\\\" [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227125 7098 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 19:15:43.227139 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 19:15:43.227162 7098 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 19:15:43.227206 7098 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:15:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbkht\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dbbzx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.280964 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b536537c-e8e6-420a-82d9-94bf662071de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e09a9bc44a05752b61d97b533117ac7aaa3c54402200551d97686b51483773b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09f4c067a1cd5938a6996ba7544dc1f3ba8b29dc495c9c07023609a97ed1e1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3ba89c0531d1af2d48c4a519d1ccf22fb8ea2fbc47cb260f5ccb497c709479b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b9c3eee1090149ebe1f47a9a9b6d2b5255b49777d8bf5ca24922b07d6e075b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44b73774db5851b969b84416d8cc494f1a85583b70b784f25b0cfdd5cc3369be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T19:14:41Z\\\",\\\"message\\\":\\\"client-ca-file\\\\nI1124 19:14:41.192462 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1124 19:14:41.192534 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1124 19:14:41.192606 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1124 19:14:41.192667 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1124 19:14:41.192682 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1124 19:14:41.192775 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\"\\\\nI1124 19:14:41.192829 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-781405056/tls.crt::/tmp/serving-cert-781405056/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764011665\\\\\\\\\\\\\\\" (2025-11-24 19:14:24 +0000 UTC to 2025-12-24 19:14:25 +0000 UTC (now=2025-11-24 19:14:41.192783942 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193042 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764011666\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764011666\\\\\\\\\\\\\\\" (2025-11-24 18:14:25 +0000 UTC to 2026-11-24 18:14:25 +0000 UTC (now=2025-11-24 19:14:41.19301958 +0000 UTC))\\\\\\\"\\\\nI1124 19:14:41.193064 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1124 19:14:41.193084 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1124 19:14:41.193110 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1124 19:14:41.193209 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8b5f39ade6b1d6f9d0fcca0c6cb40bc93553332b5d78033d3326b0b69676ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://189bc4d6090735e5fd243f3f50de07b58290a181f77aecb1d4260372ac652af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.291456 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.291517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.291539 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.291572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.291595 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.297581 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.313715 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6210f4a28c1c4bff5e1814442fdb20dfa2b0b928056e986b85f704f4ca023ccc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd4372f33253e1cbfde34b4842d9cd37a5b3f5434e0ffef8a6e6e9dcd98b697\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.327771 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.350396 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d55387-5e62-4e00-ab2d-1b11b3133843\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d152850445416bfa2e6785604b84e25f3676ce0e2000225cc83b7bf35f59443c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed65b43fd95fa91dd434bfee31fe4783f633bd2ce68b7efe40f4d6e2dd58f5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e71c2f28f22111584a5ef0ece180feb8f53e2ed2497986cce10a89b4507674fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8c95781c31eb0f5eacae2ccc4e7fb5c1eeb9397b8243db3d0c6aafb63cd9d3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b5729f0a2dff4484c911d1f626a2af27a2194cb07cfa8583ade924a3a777045\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ded3d4f659d102589c30d71475cd15ae8f4f93ced6dd7b6d83f41fd5846af38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ded3d4f659d102589c30d71475cd15ae8f4f93ced6dd7b6d83f41fd5846af38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee52112ae3bbaf08412c464af26665e659b084d1c5de5682bb83c230355dec7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee52112ae3bbaf08412c464af26665e659b084d1c5de5682bb83c230355dec7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5ce0c978c76ee9ee39d008de76021a67da21691cf6b69a8da39f2ff10a94999a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ce0c978c76ee9ee39d008de76021a67da21691cf6b69a8da39f2ff10a94999a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.361690 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304b4525cec4f603b79372ccca2d12236be7892bb6f87e7c5eae42a53c468c03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.372612 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mg7qv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38b324c3-dc7d-4555-ac0d-714a9f1a40cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T19:15:36Z\\\",\\\"message\\\":\\\"2025-11-24T19:14:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1\\\\n2025-11-24T19:14:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_056b0b58-6c73-4a11-8c90-e3d7afdb58f1 to /host/opt/cni/bin/\\\\n2025-11-24T19:14:51Z [verbose] multus-daemon started\\\\n2025-11-24T19:14:51Z [verbose] Readiness Indicator file check\\\\n2025-11-24T19:15:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9m5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mg7qv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.388658 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03570c32-b3fc-4672-9dcc-619560df1077\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0d3c850cb20d9dbc02467689abd2c7cde89d68b129e4283cd52d0368d8b6c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8fc4ec8c6a2eb8fc99c7bffd78f42cbadc3e61e734902bac00886299bbc2243b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85203175d629755e1f8e777180680e22000595ea595a91f23ac4f406edf8dc45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48b1d90baa614f1a48caa663e5dd427ffc810c6200a598724275080788e2e369\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34177373edd6cea4944175df90024576721e6384f0dd632f2e89b373aaa45b66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d3f9e968067d6882d9c8164fe6649079c5f0a0b3c3d4650c21d4df06fe792af\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c39dc4676b9269912bcbaaed58c263010cbd14b43b032ce01670e9603cde13d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T19:14:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T19:14:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-drlbp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xbc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.395528 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.395566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.395579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.395596 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.395609 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.401212 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2234d0bf6c48aee3118204b4fe8d53d754b3aacaa7d4e5464b36437b7c9dace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x4qfz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nvql4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.413665 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fjhbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56bc7a88-542e-4c82-86bb-8a4537e850b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fadb9cbb2292788d8ea234c1af449df06347e059ef97fd49db6fbd2a2a093f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fjhbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.427779 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101c186b-17e8-4a94-be8e-ae2b58a3e18c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://115d608fb3895cea2aa05d7bdd372fdde0426f8c7ca3b6cdf68b37db47c3b498\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b75e94c3e9f9259645bb267f3f76846b2f53558f9107379619b319f5f05427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:15:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5zc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pqsqf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.443957 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2ad2daaf8de095fd69c4d58361d284c598775c308d0beb0f78d6bcecfa481bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.458512 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.471188 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tqwj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6b22ba0-c1c5-4d4b-a402-966e27b725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:14:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e21340c97794f0bb00da1352b249216de6df5c52366a2e7a06c41bf5333b69b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T19:14:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w4g5t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:14:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tqwj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.485728 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T19:15:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-996cs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T19:15:01Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2pnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T19:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.498516 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.498562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.498573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.498591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.498603 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.602068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.602117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.602132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.602153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.602167 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.704849 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.704912 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.704932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.704957 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.704974 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.807507 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.807576 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.807599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.807631 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.807654 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.910806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.910864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.910889 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.910922 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:02 crc kubenswrapper[5035]: I1124 19:16:02.910944 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:02Z","lastTransitionTime":"2025-11-24T19:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.013955 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.014034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.014058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.014087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.014107 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.116918 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.116995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.117008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.117025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.117037 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.199576 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.199629 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:03 crc kubenswrapper[5035]: E1124 19:16:03.199728 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:03 crc kubenswrapper[5035]: E1124 19:16:03.199833 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.223571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.223925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.224025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.224182 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.224360 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.327440 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.327814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.328020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.328190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.328392 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.431955 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.432027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.432052 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.432082 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.432103 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.534742 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.534784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.534793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.534807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.534816 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.638461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.638510 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.638522 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.638541 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.638557 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.741692 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.741757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.741781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.741811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.741835 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.844910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.844988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.845010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.845040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.845057 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.948654 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.948711 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.948728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.948752 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:03 crc kubenswrapper[5035]: I1124 19:16:03.948769 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:03Z","lastTransitionTime":"2025-11-24T19:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.052264 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.052378 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.052395 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.052420 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.052436 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.155391 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.155466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.155486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.155510 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.155529 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.199216 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:04 crc kubenswrapper[5035]: E1124 19:16:04.199438 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.199474 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:04 crc kubenswrapper[5035]: E1124 19:16:04.199766 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.257831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.257881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.257896 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.257939 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.257955 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.360902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.361343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.361530 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.361732 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.361868 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.464831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.464871 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.464888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.464913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.464932 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.568253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.568376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.568399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.568427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.568449 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.671546 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.671615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.671640 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.671670 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.671690 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.774831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.774898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.774918 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.774946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.774968 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.877933 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.878264 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.878582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.878809 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.879026 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.982245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.982328 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.982348 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.982371 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:04 crc kubenswrapper[5035]: I1124 19:16:04.982389 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:04Z","lastTransitionTime":"2025-11-24T19:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.085637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.085688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.085705 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.085728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.085745 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.187915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.188130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.188198 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.188257 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.188346 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.199789 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.199804 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:05 crc kubenswrapper[5035]: E1124 19:16:05.199882 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:05 crc kubenswrapper[5035]: E1124 19:16:05.199976 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.291443 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.291510 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.291529 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.291554 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.291574 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.342014 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:05 crc kubenswrapper[5035]: E1124 19:16:05.342182 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:16:05 crc kubenswrapper[5035]: E1124 19:16:05.342535 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs podName:b6ac9917-ee3c-4686-a9e9-a8ba98efea02 nodeName:}" failed. No retries permitted until 2025-11-24 19:17:09.342515657 +0000 UTC m=+167.865021924 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs") pod "network-metrics-daemon-l2pnt" (UID: "b6ac9917-ee3c-4686-a9e9-a8ba98efea02") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.393563 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.393597 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.393608 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.393623 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.393634 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.496253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.496345 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.496361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.496381 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.496395 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.599662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.599734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.599748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.599774 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.599789 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.702787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.702857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.702882 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.702917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.702939 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.811615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.811668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.811681 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.811700 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.811714 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.915107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.915166 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.915185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.915207 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:05 crc kubenswrapper[5035]: I1124 19:16:05.915226 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:05Z","lastTransitionTime":"2025-11-24T19:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.018639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.018701 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.018717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.018753 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.018773 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.121746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.121818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.121840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.121866 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.121883 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.199395 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.199415 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:06 crc kubenswrapper[5035]: E1124 19:16:06.199561 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:06 crc kubenswrapper[5035]: E1124 19:16:06.199679 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.224825 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.224883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.224901 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.224966 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.224992 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.328034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.328087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.328103 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.328124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.328138 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.431630 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.431710 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.431726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.431752 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.431768 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.534732 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.534771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.534781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.534794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.534802 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.638250 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.638337 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.638357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.638381 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.638398 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.741962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.742015 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.742028 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.742046 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.742055 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.845458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.845547 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.845563 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.845586 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.845604 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.948383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.948451 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.948469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.948496 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:06 crc kubenswrapper[5035]: I1124 19:16:06.948515 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:06Z","lastTransitionTime":"2025-11-24T19:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.051786 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.051854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.051873 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.051897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.051913 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.153760 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.153797 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.153807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.153823 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.153834 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.199380 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.199463 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:07 crc kubenswrapper[5035]: E1124 19:16:07.199592 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:07 crc kubenswrapper[5035]: E1124 19:16:07.199772 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.255800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.255860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.255883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.255916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.255940 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.358087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.358231 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.358263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.358335 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.358361 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.460873 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.460942 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.460965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.460996 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.461017 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.565243 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.565398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.565422 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.565450 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.565470 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.667586 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.667636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.667648 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.667664 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.667675 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.770668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.770733 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.770750 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.770772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.770789 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.873992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.874034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.874048 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.874066 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.874080 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.976506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.976644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.976669 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.976694 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:07 crc kubenswrapper[5035]: I1124 19:16:07.976713 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:07Z","lastTransitionTime":"2025-11-24T19:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.079586 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.079646 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.079663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.079686 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.079703 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.181899 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.181944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.181956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.181975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.181988 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.199671 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.199731 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:08 crc kubenswrapper[5035]: E1124 19:16:08.199843 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:08 crc kubenswrapper[5035]: E1124 19:16:08.200069 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.284594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.284636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.284648 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.284666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.284680 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.391286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.391396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.391416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.391451 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.391470 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.494116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.494169 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.494180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.494195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.494205 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.596848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.596890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.596900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.596917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.596929 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.698982 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.699017 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.699028 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.699044 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.699088 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.800602 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.800668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.800680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.800705 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.800716 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.902853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.902891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.902900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.902915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:08 crc kubenswrapper[5035]: I1124 19:16:08.902924 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:08Z","lastTransitionTime":"2025-11-24T19:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.005162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.005203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.005216 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.005232 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.005245 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.107988 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.108027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.108035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.108048 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.108059 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.199781 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.199858 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:09 crc kubenswrapper[5035]: E1124 19:16:09.199955 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:09 crc kubenswrapper[5035]: E1124 19:16:09.200251 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.210971 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.211046 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.211110 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.211138 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.211156 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.314328 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.314365 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.314380 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.314399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.314414 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.417065 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.417140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.417165 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.417198 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.417221 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.520449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.520494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.520506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.520524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.520535 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.623572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.623639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.623660 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.623695 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.623718 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.726130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.726183 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.726198 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.726223 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.726240 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.829463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.829522 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.829540 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.829564 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.829582 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.932221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.932283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.932316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.932339 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:09 crc kubenswrapper[5035]: I1124 19:16:09.932357 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:09Z","lastTransitionTime":"2025-11-24T19:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.035636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.035679 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.035690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.035708 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.035718 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:10Z","lastTransitionTime":"2025-11-24T19:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.138343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.138461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.138475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.138501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.138516 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:10Z","lastTransitionTime":"2025-11-24T19:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.200456 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.200570 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:10 crc kubenswrapper[5035]: E1124 19:16:10.200594 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:10 crc kubenswrapper[5035]: E1124 19:16:10.200789 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.242202 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.242278 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.242393 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.242425 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.242446 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:10Z","lastTransitionTime":"2025-11-24T19:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.345581 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.345633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.345642 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.345657 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.345667 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:10Z","lastTransitionTime":"2025-11-24T19:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.448350 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.448389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.448398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.448415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.448425 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:10Z","lastTransitionTime":"2025-11-24T19:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.525620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.525659 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.525667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.525682 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.525692 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T19:16:10Z","lastTransitionTime":"2025-11-24T19:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.599788 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8"] Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.600118 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.602147 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.604212 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.604421 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.604510 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.651337 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.651314031 podStartE2EDuration="1m29.651314031s" podCreationTimestamp="2025-11-24 19:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.634664545 +0000 UTC m=+109.157170812" watchObservedRunningTime="2025-11-24 19:16:10.651314031 +0000 UTC m=+109.173820298" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.651687 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.65167922 podStartE2EDuration="56.65167922s" podCreationTimestamp="2025-11-24 19:15:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.651478245 +0000 UTC m=+109.173984502" watchObservedRunningTime="2025-11-24 19:16:10.65167922 +0000 UTC m=+109.174185487" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.692881 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=37.692853981 podStartE2EDuration="37.692853981s" podCreationTimestamp="2025-11-24 19:15:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.666959038 +0000 UTC m=+109.189465305" watchObservedRunningTime="2025-11-24 19:16:10.692853981 +0000 UTC m=+109.215360268" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.707007 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9456b-7354-4238-9d95-32c7a6fd8b55-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.707166 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9456b-7354-4238-9d95-32c7a6fd8b55-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.707346 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6fc9456b-7354-4238-9d95-32c7a6fd8b55-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.707402 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fc9456b-7354-4238-9d95-32c7a6fd8b55-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.707442 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6fc9456b-7354-4238-9d95-32c7a6fd8b55-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.750645 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.750624745 podStartE2EDuration="1m29.750624745s" podCreationTimestamp="2025-11-24 19:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.731625287 +0000 UTC m=+109.254131554" watchObservedRunningTime="2025-11-24 19:16:10.750624745 +0000 UTC m=+109.273130992" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.794853 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-fjhbz" podStartSLOduration=83.794822395 podStartE2EDuration="1m23.794822395s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.794323702 +0000 UTC m=+109.316829979" watchObservedRunningTime="2025-11-24 19:16:10.794822395 +0000 UTC m=+109.317328702" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.807247 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pqsqf" podStartSLOduration=83.807227337 podStartE2EDuration="1m23.807227337s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.80694982 +0000 UTC m=+109.329456087" watchObservedRunningTime="2025-11-24 19:16:10.807227337 +0000 UTC m=+109.329733594" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.808220 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9456b-7354-4238-9d95-32c7a6fd8b55-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.808284 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9456b-7354-4238-9d95-32c7a6fd8b55-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.808390 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6fc9456b-7354-4238-9d95-32c7a6fd8b55-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.808432 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fc9456b-7354-4238-9d95-32c7a6fd8b55-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.808495 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6fc9456b-7354-4238-9d95-32c7a6fd8b55-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.808558 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6fc9456b-7354-4238-9d95-32c7a6fd8b55-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.808592 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6fc9456b-7354-4238-9d95-32c7a6fd8b55-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.809740 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fc9456b-7354-4238-9d95-32c7a6fd8b55-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.817167 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc9456b-7354-4238-9d95-32c7a6fd8b55-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.825427 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6fc9456b-7354-4238-9d95-32c7a6fd8b55-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2mlj8\" (UID: \"6fc9456b-7354-4238-9d95-32c7a6fd8b55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.842176 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.842154401 podStartE2EDuration="9.842154401s" podCreationTimestamp="2025-11-24 19:16:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.841722279 +0000 UTC m=+109.364228546" watchObservedRunningTime="2025-11-24 19:16:10.842154401 +0000 UTC m=+109.364660658" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.869860 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-mg7qv" podStartSLOduration=83.869840931 podStartE2EDuration="1m23.869840931s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.86943417 +0000 UTC m=+109.391940437" watchObservedRunningTime="2025-11-24 19:16:10.869840931 +0000 UTC m=+109.392347188" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.888723 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xbc8w" podStartSLOduration=83.888703024 podStartE2EDuration="1m23.888703024s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.887996046 +0000 UTC m=+109.410502303" watchObservedRunningTime="2025-11-24 19:16:10.888703024 +0000 UTC m=+109.411209281" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.922608 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podStartSLOduration=83.922591481 podStartE2EDuration="1m23.922591481s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.904136217 +0000 UTC m=+109.426642474" watchObservedRunningTime="2025-11-24 19:16:10.922591481 +0000 UTC m=+109.445097738" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.923062 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" Nov 24 19:16:10 crc kubenswrapper[5035]: I1124 19:16:10.959566 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-tqwj5" podStartSLOduration=83.959545808 podStartE2EDuration="1m23.959545808s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:10.958375217 +0000 UTC m=+109.480881484" watchObservedRunningTime="2025-11-24 19:16:10.959545808 +0000 UTC m=+109.482052065" Nov 24 19:16:11 crc kubenswrapper[5035]: I1124 19:16:11.192399 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" event={"ID":"6fc9456b-7354-4238-9d95-32c7a6fd8b55","Type":"ContainerStarted","Data":"68e69e77da91c2fec2f58811c6adedd8e4cd419dcb390cafbb66dec9831e93ec"} Nov 24 19:16:11 crc kubenswrapper[5035]: I1124 19:16:11.192497 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" event={"ID":"6fc9456b-7354-4238-9d95-32c7a6fd8b55","Type":"ContainerStarted","Data":"4407e491d3e655d9e65bbe0e44018da827dc0a6272bfd316ed751492ea826401"} Nov 24 19:16:11 crc kubenswrapper[5035]: I1124 19:16:11.200012 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:11 crc kubenswrapper[5035]: E1124 19:16:11.200239 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:11 crc kubenswrapper[5035]: I1124 19:16:11.200636 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:11 crc kubenswrapper[5035]: E1124 19:16:11.200763 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:12 crc kubenswrapper[5035]: I1124 19:16:12.199199 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:12 crc kubenswrapper[5035]: I1124 19:16:12.199256 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:12 crc kubenswrapper[5035]: E1124 19:16:12.201044 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:12 crc kubenswrapper[5035]: E1124 19:16:12.201162 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:13 crc kubenswrapper[5035]: I1124 19:16:13.199237 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:13 crc kubenswrapper[5035]: I1124 19:16:13.199260 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:13 crc kubenswrapper[5035]: E1124 19:16:13.199786 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:13 crc kubenswrapper[5035]: E1124 19:16:13.199878 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:14 crc kubenswrapper[5035]: I1124 19:16:14.200787 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:14 crc kubenswrapper[5035]: I1124 19:16:14.200967 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:14 crc kubenswrapper[5035]: E1124 19:16:14.201105 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:14 crc kubenswrapper[5035]: E1124 19:16:14.201176 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:15 crc kubenswrapper[5035]: I1124 19:16:15.199598 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:15 crc kubenswrapper[5035]: I1124 19:16:15.199643 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:15 crc kubenswrapper[5035]: E1124 19:16:15.199777 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:15 crc kubenswrapper[5035]: E1124 19:16:15.200003 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:15 crc kubenswrapper[5035]: I1124 19:16:15.201473 5035 scope.go:117] "RemoveContainer" containerID="6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d" Nov 24 19:16:15 crc kubenswrapper[5035]: E1124 19:16:15.201775 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dbbzx_openshift-ovn-kubernetes(4d9cb31b-5f2a-4594-ad12-718b6e99d15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" Nov 24 19:16:16 crc kubenswrapper[5035]: I1124 19:16:16.200129 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:16 crc kubenswrapper[5035]: I1124 19:16:16.200462 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:16 crc kubenswrapper[5035]: E1124 19:16:16.200601 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:16 crc kubenswrapper[5035]: E1124 19:16:16.200819 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:17 crc kubenswrapper[5035]: I1124 19:16:17.199996 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:17 crc kubenswrapper[5035]: I1124 19:16:17.200013 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:17 crc kubenswrapper[5035]: E1124 19:16:17.200175 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:17 crc kubenswrapper[5035]: E1124 19:16:17.200282 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:18 crc kubenswrapper[5035]: I1124 19:16:18.199095 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:18 crc kubenswrapper[5035]: E1124 19:16:18.199220 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:18 crc kubenswrapper[5035]: I1124 19:16:18.199266 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:18 crc kubenswrapper[5035]: E1124 19:16:18.199576 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:19 crc kubenswrapper[5035]: I1124 19:16:19.200082 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:19 crc kubenswrapper[5035]: I1124 19:16:19.200095 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:19 crc kubenswrapper[5035]: E1124 19:16:19.200279 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:19 crc kubenswrapper[5035]: E1124 19:16:19.200396 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:20 crc kubenswrapper[5035]: I1124 19:16:20.200148 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:20 crc kubenswrapper[5035]: I1124 19:16:20.200365 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:20 crc kubenswrapper[5035]: E1124 19:16:20.201223 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:20 crc kubenswrapper[5035]: E1124 19:16:20.201082 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:21 crc kubenswrapper[5035]: I1124 19:16:21.199799 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:21 crc kubenswrapper[5035]: E1124 19:16:21.199978 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:21 crc kubenswrapper[5035]: I1124 19:16:21.200263 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:21 crc kubenswrapper[5035]: E1124 19:16:21.201133 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:22 crc kubenswrapper[5035]: I1124 19:16:22.199941 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:22 crc kubenswrapper[5035]: I1124 19:16:22.200074 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:22 crc kubenswrapper[5035]: E1124 19:16:22.200675 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:22 crc kubenswrapper[5035]: E1124 19:16:22.200830 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:22 crc kubenswrapper[5035]: E1124 19:16:22.214680 5035 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 19:16:22 crc kubenswrapper[5035]: E1124 19:16:22.438162 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 19:16:22 crc kubenswrapper[5035]: E1124 19:16:22.919148 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38b324c3_dc7d_4555_ac0d_714a9f1a40cf.slice/crio-conmon-91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965.scope\": RecentStats: unable to find data in memory cache]" Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.199636 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.199643 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:23 crc kubenswrapper[5035]: E1124 19:16:23.199939 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:23 crc kubenswrapper[5035]: E1124 19:16:23.199798 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.232438 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/1.log" Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.233243 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/0.log" Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.233348 5035 generic.go:334] "Generic (PLEG): container finished" podID="38b324c3-dc7d-4555-ac0d-714a9f1a40cf" containerID="91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965" exitCode=1 Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.233401 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerDied","Data":"91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965"} Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.233472 5035 scope.go:117] "RemoveContainer" containerID="6863b6232550398da18bed5fdd3ab50be6f3dca7ef840a6827268361c649a88c" Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.234108 5035 scope.go:117] "RemoveContainer" containerID="91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965" Nov 24 19:16:23 crc kubenswrapper[5035]: E1124 19:16:23.234636 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-mg7qv_openshift-multus(38b324c3-dc7d-4555-ac0d-714a9f1a40cf)\"" pod="openshift-multus/multus-mg7qv" podUID="38b324c3-dc7d-4555-ac0d-714a9f1a40cf" Nov 24 19:16:23 crc kubenswrapper[5035]: I1124 19:16:23.264134 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mlj8" podStartSLOduration=96.264105739 podStartE2EDuration="1m36.264105739s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:11.206664122 +0000 UTC m=+109.729170379" watchObservedRunningTime="2025-11-24 19:16:23.264105739 +0000 UTC m=+121.786612036" Nov 24 19:16:24 crc kubenswrapper[5035]: I1124 19:16:24.200034 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:24 crc kubenswrapper[5035]: E1124 19:16:24.200179 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:24 crc kubenswrapper[5035]: I1124 19:16:24.200407 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:24 crc kubenswrapper[5035]: E1124 19:16:24.200616 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:24 crc kubenswrapper[5035]: I1124 19:16:24.238971 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/1.log" Nov 24 19:16:25 crc kubenswrapper[5035]: I1124 19:16:25.199706 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:25 crc kubenswrapper[5035]: E1124 19:16:25.199831 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:25 crc kubenswrapper[5035]: I1124 19:16:25.199717 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:25 crc kubenswrapper[5035]: E1124 19:16:25.200032 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:26 crc kubenswrapper[5035]: I1124 19:16:26.199993 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:26 crc kubenswrapper[5035]: E1124 19:16:26.200130 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:26 crc kubenswrapper[5035]: I1124 19:16:26.200200 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:26 crc kubenswrapper[5035]: E1124 19:16:26.200475 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:27 crc kubenswrapper[5035]: I1124 19:16:27.199621 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:27 crc kubenswrapper[5035]: I1124 19:16:27.199636 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:27 crc kubenswrapper[5035]: E1124 19:16:27.199821 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:27 crc kubenswrapper[5035]: E1124 19:16:27.200049 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:27 crc kubenswrapper[5035]: I1124 19:16:27.201080 5035 scope.go:117] "RemoveContainer" containerID="6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d" Nov 24 19:16:27 crc kubenswrapper[5035]: E1124 19:16:27.439602 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.199826 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.199838 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:28 crc kubenswrapper[5035]: E1124 19:16:28.199956 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:28 crc kubenswrapper[5035]: E1124 19:16:28.200064 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.254115 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-l2pnt"] Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.254244 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:28 crc kubenswrapper[5035]: E1124 19:16:28.254418 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.261658 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/3.log" Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.264224 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerStarted","Data":"e8ead8c9c31b17a0cb024c98dd49f1f9c8366067fc37af3c460a0a2e98a35619"} Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.280107 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:16:28 crc kubenswrapper[5035]: I1124 19:16:28.314661 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podStartSLOduration=101.314636439 podStartE2EDuration="1m41.314636439s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:28.314376462 +0000 UTC m=+126.836882749" watchObservedRunningTime="2025-11-24 19:16:28.314636439 +0000 UTC m=+126.837142726" Nov 24 19:16:29 crc kubenswrapper[5035]: I1124 19:16:29.199658 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:29 crc kubenswrapper[5035]: E1124 19:16:29.199892 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:30 crc kubenswrapper[5035]: I1124 19:16:30.199684 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:30 crc kubenswrapper[5035]: E1124 19:16:30.199856 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:30 crc kubenswrapper[5035]: I1124 19:16:30.199967 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:30 crc kubenswrapper[5035]: I1124 19:16:30.199988 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:30 crc kubenswrapper[5035]: E1124 19:16:30.200236 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:30 crc kubenswrapper[5035]: E1124 19:16:30.200341 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:31 crc kubenswrapper[5035]: I1124 19:16:31.199635 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:31 crc kubenswrapper[5035]: E1124 19:16:31.199836 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:32 crc kubenswrapper[5035]: I1124 19:16:32.199989 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:32 crc kubenswrapper[5035]: E1124 19:16:32.201844 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:32 crc kubenswrapper[5035]: I1124 19:16:32.201937 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:32 crc kubenswrapper[5035]: I1124 19:16:32.201937 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:32 crc kubenswrapper[5035]: E1124 19:16:32.202099 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:32 crc kubenswrapper[5035]: E1124 19:16:32.202249 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:32 crc kubenswrapper[5035]: E1124 19:16:32.441805 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 19:16:33 crc kubenswrapper[5035]: I1124 19:16:33.199041 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:33 crc kubenswrapper[5035]: E1124 19:16:33.199551 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:34 crc kubenswrapper[5035]: I1124 19:16:34.199081 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:34 crc kubenswrapper[5035]: I1124 19:16:34.200168 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:34 crc kubenswrapper[5035]: E1124 19:16:34.200473 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:34 crc kubenswrapper[5035]: I1124 19:16:34.200517 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:34 crc kubenswrapper[5035]: E1124 19:16:34.200671 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:34 crc kubenswrapper[5035]: E1124 19:16:34.200769 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:35 crc kubenswrapper[5035]: I1124 19:16:35.200005 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:35 crc kubenswrapper[5035]: E1124 19:16:35.200133 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:36 crc kubenswrapper[5035]: I1124 19:16:36.199219 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:36 crc kubenswrapper[5035]: I1124 19:16:36.199443 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:36 crc kubenswrapper[5035]: I1124 19:16:36.199346 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:36 crc kubenswrapper[5035]: E1124 19:16:36.199514 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:36 crc kubenswrapper[5035]: E1124 19:16:36.199674 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:36 crc kubenswrapper[5035]: E1124 19:16:36.199913 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:37 crc kubenswrapper[5035]: I1124 19:16:37.199976 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:37 crc kubenswrapper[5035]: E1124 19:16:37.200152 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:37 crc kubenswrapper[5035]: I1124 19:16:37.201504 5035 scope.go:117] "RemoveContainer" containerID="91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965" Nov 24 19:16:37 crc kubenswrapper[5035]: E1124 19:16:37.443128 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 19:16:38 crc kubenswrapper[5035]: I1124 19:16:38.199947 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:38 crc kubenswrapper[5035]: I1124 19:16:38.200079 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:38 crc kubenswrapper[5035]: E1124 19:16:38.200186 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:38 crc kubenswrapper[5035]: I1124 19:16:38.200219 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:38 crc kubenswrapper[5035]: E1124 19:16:38.200395 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:38 crc kubenswrapper[5035]: E1124 19:16:38.200456 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:38 crc kubenswrapper[5035]: I1124 19:16:38.308364 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/1.log" Nov 24 19:16:38 crc kubenswrapper[5035]: I1124 19:16:38.308426 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerStarted","Data":"01506ee8a33f0daf99578100ec1724d23ab8e5d245387a0430b41fa3e72a0a19"} Nov 24 19:16:39 crc kubenswrapper[5035]: I1124 19:16:39.199847 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:39 crc kubenswrapper[5035]: E1124 19:16:39.200095 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:40 crc kubenswrapper[5035]: I1124 19:16:40.199244 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:40 crc kubenswrapper[5035]: I1124 19:16:40.199287 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:40 crc kubenswrapper[5035]: I1124 19:16:40.199359 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:40 crc kubenswrapper[5035]: E1124 19:16:40.199409 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:40 crc kubenswrapper[5035]: E1124 19:16:40.199513 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:40 crc kubenswrapper[5035]: E1124 19:16:40.199623 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:41 crc kubenswrapper[5035]: I1124 19:16:41.199919 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:41 crc kubenswrapper[5035]: E1124 19:16:41.200057 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 19:16:42 crc kubenswrapper[5035]: I1124 19:16:42.199766 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:42 crc kubenswrapper[5035]: I1124 19:16:42.199862 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:42 crc kubenswrapper[5035]: I1124 19:16:42.199899 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:42 crc kubenswrapper[5035]: E1124 19:16:42.203666 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2pnt" podUID="b6ac9917-ee3c-4686-a9e9-a8ba98efea02" Nov 24 19:16:42 crc kubenswrapper[5035]: E1124 19:16:42.204439 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 19:16:42 crc kubenswrapper[5035]: E1124 19:16:42.204822 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 19:16:42 crc kubenswrapper[5035]: I1124 19:16:42.565121 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:16:43 crc kubenswrapper[5035]: I1124 19:16:43.199222 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:43 crc kubenswrapper[5035]: I1124 19:16:43.201600 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 19:16:43 crc kubenswrapper[5035]: I1124 19:16:43.203252 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 19:16:44 crc kubenswrapper[5035]: I1124 19:16:44.200002 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:44 crc kubenswrapper[5035]: I1124 19:16:44.200062 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:16:44 crc kubenswrapper[5035]: I1124 19:16:44.200203 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:44 crc kubenswrapper[5035]: I1124 19:16:44.202875 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 19:16:44 crc kubenswrapper[5035]: I1124 19:16:44.204497 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 19:16:44 crc kubenswrapper[5035]: I1124 19:16:44.205571 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 19:16:44 crc kubenswrapper[5035]: I1124 19:16:44.205776 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.024791 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.024840 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.024867 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.024890 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.025949 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.030872 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.031973 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.033430 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.119450 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.125840 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:50 crc kubenswrapper[5035]: E1124 19:16:50.126062 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:18:52.126029557 +0000 UTC m=+270.648535844 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.226707 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 19:16:50 crc kubenswrapper[5035]: I1124 19:16:50.242761 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 19:16:50 crc kubenswrapper[5035]: W1124 19:16:50.455661 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-1c2a870d61f16d87a61b1a7520a4b3aa5e0b42dafebbd3f8ade36c5b7f35d832 WatchSource:0}: Error finding container 1c2a870d61f16d87a61b1a7520a4b3aa5e0b42dafebbd3f8ade36c5b7f35d832: Status 404 returned error can't find the container with id 1c2a870d61f16d87a61b1a7520a4b3aa5e0b42dafebbd3f8ade36c5b7f35d832 Nov 24 19:16:50 crc kubenswrapper[5035]: W1124 19:16:50.478256 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-a073537da2880c5fba3259ede463dd08022ba7dcfaf849c12592a2a825fbd571 WatchSource:0}: Error finding container a073537da2880c5fba3259ede463dd08022ba7dcfaf849c12592a2a825fbd571: Status 404 returned error can't find the container with id a073537da2880c5fba3259ede463dd08022ba7dcfaf849c12592a2a825fbd571 Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.217249 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.269717 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lwfhz"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.270458 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.272053 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.272637 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.272841 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.274362 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r2vkf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.275056 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.275893 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.276254 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.276360 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.276646 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.276766 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5z8h5"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.276973 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.277861 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.282870 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.283144 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.283209 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.283430 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.284517 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.285994 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.287428 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.288220 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.289002 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qfxzs"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.289732 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.290854 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtsbg"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.291563 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.291926 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.292101 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.292254 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.292880 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.293471 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.293637 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.294123 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.294339 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.294420 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.294658 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.294861 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.295693 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.294886 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.294942 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.295511 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.297093 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8b8fg"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.297731 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.299560 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300083 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300173 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-v8j4d"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300436 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300600 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300717 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300855 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300934 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-sblll"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300973 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.300978 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-v8j4d" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.301148 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.313649 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.314730 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.315674 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.316208 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.316377 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.316985 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.318021 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-58jzg"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.319252 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.319457 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.320056 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.320601 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.365570 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.366003 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.366644 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.369286 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v5t7x"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.370121 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.371567 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.372837 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.372928 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373044 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373113 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373178 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373244 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373332 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373373 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373429 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373508 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373585 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.373775 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-audit-policies\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.374118 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdq8n\" (UniqueName: \"kubernetes.io/projected/03ba0026-5797-472e-a49c-9789c64cb162-kube-api-access-pdq8n\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.374795 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.374953 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375382 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375410 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-encryption-config\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375432 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03ba0026-5797-472e-a49c-9789c64cb162-audit-dir\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375467 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-etcd-client\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375486 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-serving-cert\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375502 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375698 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.375918 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.376266 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5z8h5"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.376605 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.378726 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"de64fcfd29a42d234402f29384f1344878859eb3fd58c63851f2a06e5d556deb"} Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.378766 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1c2a870d61f16d87a61b1a7520a4b3aa5e0b42dafebbd3f8ade36c5b7f35d832"} Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.384923 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.385706 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.394259 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"34de71b363ff129db46d75ba862a8acf15eaa11bea3eb03875227f2cff46492c"} Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.394320 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a073537da2880c5fba3259ede463dd08022ba7dcfaf849c12592a2a825fbd571"} Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.397349 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f3eb43afdff20d497f89797b2bc1dbec6fe68716c51a2df8289869de89afcac0"} Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.397421 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"96c89ba56be0d4505b2778f4d8b904e195b77779b5460426e85622398307811e"} Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.398068 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.429787 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.429915 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.430037 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.430466 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.430507 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.430651 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.431162 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.432785 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.436178 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r2vkf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.436278 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.436477 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.436725 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.436866 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.437007 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.437137 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.437245 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.437396 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.437509 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.437629 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.437895 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.438121 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.439141 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.441098 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.441282 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.441420 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.441530 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.441714 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.441888 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.442002 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.442023 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.442119 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.442459 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.442536 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.442608 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.442830 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.450112 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.450483 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.450678 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.450718 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.450797 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.452531 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.452711 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fk6tx"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.453153 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.455403 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.455777 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.455845 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.456072 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.457138 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.457204 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.457682 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.457833 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.474026 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.474810 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.477639 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.480830 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-oauth-config\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.480880 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.480963 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-etcd-serving-ca\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.480998 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-serving-cert\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481021 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/236d3c3c-fe73-409f-b39c-7bb3ba97c902-images\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481111 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-client-ca\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481134 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf257\" (UniqueName: \"kubernetes.io/projected/bb15b826-35be-4433-9e0f-88e013e8141d-kube-api-access-xf257\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481208 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481232 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-etcd-client\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481261 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86c7681b-593c-4e81-8df4-0e90bab1facd-serving-cert\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481277 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc4c4\" (UniqueName: \"kubernetes.io/projected/727a0b65-6c8b-477d-8743-e3bd88a7c904-kube-api-access-kc4c4\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481398 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-encryption-config\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.481425 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03ba0026-5797-472e-a49c-9789c64cb162-audit-dir\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.482360 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.482988 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.489480 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.489544 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.489818 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.489849 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.491120 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.491369 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.491640 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-policies\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.491705 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.491952 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03ba0026-5797-472e-a49c-9789c64cb162-audit-dir\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492004 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-image-import-ca\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492109 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/170f65fd-91b4-46ef-a97e-dfe96117d8b8-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gl4nm\" (UID: \"170f65fd-91b4-46ef-a97e-dfe96117d8b8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492206 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492269 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492326 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-console-config\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492363 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492421 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-989gg\" (UniqueName: \"kubernetes.io/projected/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-kube-api-access-989gg\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492459 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-audit-dir\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492553 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-etcd-client\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492595 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-config\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492621 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmvfz\" (UniqueName: \"kubernetes.io/projected/1d176339-2edc-4cf3-ad85-f4a0519f41c2-kube-api-access-zmvfz\") pod \"downloads-7954f5f757-v8j4d\" (UID: \"1d176339-2edc-4cf3-ad85-f4a0519f41c2\") " pod="openshift-console/downloads-7954f5f757-v8j4d" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492647 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236d3c3c-fe73-409f-b39c-7bb3ba97c902-config\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492670 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x875\" (UniqueName: \"kubernetes.io/projected/236d3c3c-fe73-409f-b39c-7bb3ba97c902-kube-api-access-4x875\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492793 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492936 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pknp2\" (UniqueName: \"kubernetes.io/projected/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-kube-api-access-pknp2\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.492980 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493070 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493115 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493220 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb15b826-35be-4433-9e0f-88e013e8141d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493249 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50060f5-1b00-483a-914b-c34277882c04-serving-cert\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493389 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2z6v\" (UniqueName: \"kubernetes.io/projected/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-kube-api-access-p2z6v\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493473 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-audit-policies\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493503 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-service-ca\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493547 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdq8n\" (UniqueName: \"kubernetes.io/projected/03ba0026-5797-472e-a49c-9789c64cb162-kube-api-access-pdq8n\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493577 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-trusted-ca-bundle\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493597 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/236d3c3c-fe73-409f-b39c-7bb3ba97c902-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493627 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-oauth-serving-cert\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493651 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-config\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493627 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493673 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-encryption-config\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493718 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hw8n\" (UniqueName: \"kubernetes.io/projected/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-kube-api-access-5hw8n\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-audit\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493758 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493794 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-serving-cert\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493817 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727a0b65-6c8b-477d-8743-e3bd88a7c904-serving-cert\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493833 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-serving-cert\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493856 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-868kx\" (UniqueName: \"kubernetes.io/projected/d50060f5-1b00-483a-914b-c34277882c04-kube-api-access-868kx\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493872 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493889 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493906 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnczm\" (UniqueName: \"kubernetes.io/projected/170f65fd-91b4-46ef-a97e-dfe96117d8b8-kube-api-access-lnczm\") pod \"cluster-samples-operator-665b6dd947-gl4nm\" (UID: \"170f65fd-91b4-46ef-a97e-dfe96117d8b8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493922 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-dir\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493944 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493967 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.493988 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-machine-approver-tls\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494017 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494038 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-config\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494057 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494076 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-trusted-ca\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494096 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-audit-policies\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494105 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494130 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-node-pullsecrets\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494158 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494180 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494205 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-config\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494230 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494260 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5wjh\" (UniqueName: \"kubernetes.io/projected/2a0d45af-2aba-4830-a849-c2933f94fb83-kube-api-access-z5wjh\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494284 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb15b826-35be-4433-9e0f-88e013e8141d-serving-cert\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494359 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-config\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494401 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-serving-cert\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494429 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhn2c\" (UniqueName: \"kubernetes.io/projected/fb6648af-1ab7-419e-bdb3-8f696aaabd80-kube-api-access-qhn2c\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494449 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p76v\" (UniqueName: \"kubernetes.io/projected/86c7681b-593c-4e81-8df4-0e90bab1facd-kube-api-access-8p76v\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494466 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-config\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494483 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-client-ca\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494501 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-auth-proxy-config\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.494407 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03ba0026-5797-472e-a49c-9789c64cb162-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.495884 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.496520 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.497334 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.498800 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.498954 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-etcd-client\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.502334 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.502773 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zfmld"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.503251 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q2jhw"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.503386 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.503545 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-serving-cert\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.503710 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.503891 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.505447 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.505738 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.506186 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.506520 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.506592 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.506920 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.507011 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.507189 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.508317 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.508984 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.509445 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zv4sz"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.509950 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.511657 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.512227 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.514705 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.515546 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.516246 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lwfhz"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.516895 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.517844 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03ba0026-5797-472e-a49c-9789c64cb162-encryption-config\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.524398 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-w8dwg"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.525162 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.525371 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.525977 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v5t7x"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.528445 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2j9k6"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.529258 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.530741 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.532631 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.534444 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.535117 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-47f49"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.535815 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-58jzg"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.535907 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.535949 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtsbg"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.536159 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.536307 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.537842 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qfxzs"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.542884 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-v8j4d"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.545103 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.548318 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.550376 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vzvml"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.553702 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.554593 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.554670 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.557352 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.562813 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.563838 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.565064 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8b8fg"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.568399 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2p6pp"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.569076 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.569178 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.569516 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.570428 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.571870 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.572904 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.574485 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.575282 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2j9k6"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.577771 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.579517 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fk6tx"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.582146 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q2jhw"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.584768 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zfmld"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.585900 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.587023 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.587938 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.590110 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vzvml"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.590480 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.591543 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-wj8q6"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.592520 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.592560 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-csmtf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.593106 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.593578 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.593954 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.594993 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b08695f9-1ea8-4852-8b0a-0d548f000194-metrics-tls\") pod \"dns-operator-744455d44c-v5t7x\" (UID: \"b08695f9-1ea8-4852-8b0a-0d548f000194\") " pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595031 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595064 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595088 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5wjh\" (UniqueName: \"kubernetes.io/projected/2a0d45af-2aba-4830-a849-c2933f94fb83-kube-api-access-z5wjh\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595115 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrsv4\" (UniqueName: \"kubernetes.io/projected/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-kube-api-access-wrsv4\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595141 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-config\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595165 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-client-ca\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595190 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-node-bootstrap-token\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595216 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595236 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-signing-key\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595259 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/236d3c3c-fe73-409f-b39c-7bb3ba97c902-images\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595280 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-client-ca\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595321 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf257\" (UniqueName: \"kubernetes.io/projected/bb15b826-35be-4433-9e0f-88e013e8141d-kube-api-access-xf257\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595346 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595368 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-etcd-client\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595411 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86c7681b-593c-4e81-8df4-0e90bab1facd-serving-cert\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595435 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc4c4\" (UniqueName: \"kubernetes.io/projected/727a0b65-6c8b-477d-8743-e3bd88a7c904-kube-api-access-kc4c4\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595458 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-serving-cert\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595482 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c8bd5ed-e597-4095-b5e1-146db751adec-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595518 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595541 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/170f65fd-91b4-46ef-a97e-dfe96117d8b8-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gl4nm\" (UID: \"170f65fd-91b4-46ef-a97e-dfe96117d8b8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595565 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flsrf\" (UniqueName: \"kubernetes.io/projected/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-kube-api-access-flsrf\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595589 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-989gg\" (UniqueName: \"kubernetes.io/projected/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-kube-api-access-989gg\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595612 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595631 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-certs\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-audit-dir\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595680 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-srv-cert\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595704 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-config\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595729 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmvfz\" (UniqueName: \"kubernetes.io/projected/1d176339-2edc-4cf3-ad85-f4a0519f41c2-kube-api-access-zmvfz\") pod \"downloads-7954f5f757-v8j4d\" (UID: \"1d176339-2edc-4cf3-ad85-f4a0519f41c2\") " pod="openshift-console/downloads-7954f5f757-v8j4d" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595756 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x875\" (UniqueName: \"kubernetes.io/projected/236d3c3c-fe73-409f-b39c-7bb3ba97c902-kube-api-access-4x875\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595780 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-signing-cabundle\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595803 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595828 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb15b826-35be-4433-9e0f-88e013e8141d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595854 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595878 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2z6v\" (UniqueName: \"kubernetes.io/projected/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-kube-api-access-p2z6v\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595902 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50060f5-1b00-483a-914b-c34277882c04-serving-cert\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595925 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-service-ca\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595954 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-trusted-ca-bundle\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.595978 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596001 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-config\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596025 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-audit\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596050 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596075 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8bd5ed-e597-4095-b5e1-146db751adec-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596097 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-serving-cert\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596160 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596191 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c8bd5ed-e597-4095-b5e1-146db751adec-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596216 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596239 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8105a388-83c1-4372-bd51-fcbc2a2e9198-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6r9rt\" (UID: \"8105a388-83c1-4372-bd51-fcbc2a2e9198\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnczm\" (UniqueName: \"kubernetes.io/projected/170f65fd-91b4-46ef-a97e-dfe96117d8b8-kube-api-access-lnczm\") pod \"cluster-samples-operator-665b6dd947-gl4nm\" (UID: \"170f65fd-91b4-46ef-a97e-dfe96117d8b8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596309 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-proxy-tls\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596335 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-ca\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596360 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-dir\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596384 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596406 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-machine-approver-tls\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596429 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jr6b\" (UniqueName: \"kubernetes.io/projected/fa17b560-2306-4c7e-ae02-02f6f16db399-kube-api-access-6jr6b\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596452 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-trusted-ca\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596476 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-node-pullsecrets\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596499 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596522 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596546 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596569 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-config\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596596 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb15b826-35be-4433-9e0f-88e013e8141d-serving-cert\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596620 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-config\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596643 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dszmc\" (UniqueName: \"kubernetes.io/projected/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-kube-api-access-dszmc\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596666 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhn2c\" (UniqueName: \"kubernetes.io/projected/fb6648af-1ab7-419e-bdb3-8f696aaabd80-kube-api-access-qhn2c\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596688 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p76v\" (UniqueName: \"kubernetes.io/projected/86c7681b-593c-4e81-8df4-0e90bab1facd-kube-api-access-8p76v\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596714 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-auth-proxy-config\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596737 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xbv2\" (UniqueName: \"kubernetes.io/projected/9f27c073-7ed7-45e0-aa21-2a8d294e2050-kube-api-access-6xbv2\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596759 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-oauth-config\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596797 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596822 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-etcd-serving-ca\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596843 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-srv-cert\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596866 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-serving-cert\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596906 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596931 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdvwz\" (UniqueName: \"kubernetes.io/projected/b08695f9-1ea8-4852-8b0a-0d548f000194-kube-api-access-kdvwz\") pod \"dns-operator-744455d44c-v5t7x\" (UID: \"b08695f9-1ea8-4852-8b0a-0d548f000194\") " pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596959 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-image-import-ca\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.596983 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-policies\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597004 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-console-config\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597029 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597054 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597077 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjlt8\" (UniqueName: \"kubernetes.io/projected/8105a388-83c1-4372-bd51-fcbc2a2e9198-kube-api-access-rjlt8\") pod \"control-plane-machine-set-operator-78cbb6b69f-6r9rt\" (UID: \"8105a388-83c1-4372-bd51-fcbc2a2e9198\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597104 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236d3c3c-fe73-409f-b39c-7bb3ba97c902-config\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597131 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597155 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pknp2\" (UniqueName: \"kubernetes.io/projected/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-kube-api-access-pknp2\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597182 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597211 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597233 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpjmk\" (UniqueName: \"kubernetes.io/projected/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-kube-api-access-hpjmk\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597258 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597283 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-service-ca\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597324 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g7nd\" (UniqueName: \"kubernetes.io/projected/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-kube-api-access-7g7nd\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597346 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-client\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597385 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/236d3c3c-fe73-409f-b39c-7bb3ba97c902-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597409 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-config\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597430 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjghs\" (UniqueName: \"kubernetes.io/projected/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-kube-api-access-mjghs\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597454 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-encryption-config\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597479 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hw8n\" (UniqueName: \"kubernetes.io/projected/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-kube-api-access-5hw8n\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597504 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-oauth-serving-cert\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597528 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-config\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597551 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-config\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597575 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727a0b65-6c8b-477d-8743-e3bd88a7c904-serving-cert\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597598 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-serving-cert\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597624 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-868kx\" (UniqueName: \"kubernetes.io/projected/d50060f5-1b00-483a-914b-c34277882c04-kube-api-access-868kx\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597648 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597671 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs25h\" (UniqueName: \"kubernetes.io/projected/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-kube-api-access-gs25h\") pod \"ingress-canary-wj8q6\" (UID: \"146bb9f2-5eb7-43db-a96f-6588ac7bbd12\") " pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597696 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597720 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597745 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-config\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597770 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-cert\") pod \"ingress-canary-wj8q6\" (UID: \"146bb9f2-5eb7-43db-a96f-6588ac7bbd12\") " pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597793 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-profile-collector-cert\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597930 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zv4sz"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597954 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-47f49"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.597966 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-csmtf"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.598028 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-audit-dir\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.599025 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.599067 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-wj8q6"] Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.599379 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-config\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.600138 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.600412 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/170f65fd-91b4-46ef-a97e-dfe96117d8b8-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gl4nm\" (UID: \"170f65fd-91b4-46ef-a97e-dfe96117d8b8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.600689 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-dir\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.600844 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-config\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.600917 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-etcd-serving-ca\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.601364 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-policies\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.602022 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-oauth-serving-cert\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.602124 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-image-import-ca\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.602535 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-config\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.602774 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-console-config\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.603877 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-config\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.604122 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.604151 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-auth-proxy-config\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.604158 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.604540 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.604562 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/236d3c3c-fe73-409f-b39c-7bb3ba97c902-config\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.605867 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.606526 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.606544 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-trusted-ca\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.606574 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.606712 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86c7681b-593c-4e81-8df4-0e90bab1facd-service-ca-bundle\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.606823 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-client-ca\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.607052 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/236d3c3c-fe73-409f-b39c-7bb3ba97c902-images\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.607157 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-audit\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.607520 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.607696 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-config\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.607790 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.607840 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-node-pullsecrets\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.607971 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-service-ca\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.608089 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50060f5-1b00-483a-914b-c34277882c04-serving-cert\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.608301 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-config\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.608586 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-trusted-ca-bundle\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.608753 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-client-ca\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.609002 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb15b826-35be-4433-9e0f-88e013e8141d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.609081 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-serving-cert\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.609305 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.609421 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.610436 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727a0b65-6c8b-477d-8743-e3bd88a7c904-serving-cert\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.610860 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-etcd-client\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.611205 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-machine-approver-tls\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.611371 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.611470 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.611678 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-serving-cert\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.611800 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/236d3c3c-fe73-409f-b39c-7bb3ba97c902-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.612053 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-serving-cert\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.612153 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.612434 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.612682 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.612898 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-oauth-config\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.616697 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-encryption-config\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.616784 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86c7681b-593c-4e81-8df4-0e90bab1facd-serving-cert\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.617027 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb15b826-35be-4433-9e0f-88e013e8141d-serving-cert\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.633267 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.652828 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.673219 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.693034 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698230 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c8bd5ed-e597-4095-b5e1-146db751adec-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698273 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698310 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8105a388-83c1-4372-bd51-fcbc2a2e9198-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6r9rt\" (UID: \"8105a388-83c1-4372-bd51-fcbc2a2e9198\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698330 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-proxy-tls\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698347 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-ca\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698363 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jr6b\" (UniqueName: \"kubernetes.io/projected/fa17b560-2306-4c7e-ae02-02f6f16db399-kube-api-access-6jr6b\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698381 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698397 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dszmc\" (UniqueName: \"kubernetes.io/projected/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-kube-api-access-dszmc\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698437 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xbv2\" (UniqueName: \"kubernetes.io/projected/9f27c073-7ed7-45e0-aa21-2a8d294e2050-kube-api-access-6xbv2\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698453 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698473 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-srv-cert\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698489 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698506 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdvwz\" (UniqueName: \"kubernetes.io/projected/b08695f9-1ea8-4852-8b0a-0d548f000194-kube-api-access-kdvwz\") pod \"dns-operator-744455d44c-v5t7x\" (UID: \"b08695f9-1ea8-4852-8b0a-0d548f000194\") " pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698525 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjlt8\" (UniqueName: \"kubernetes.io/projected/8105a388-83c1-4372-bd51-fcbc2a2e9198-kube-api-access-rjlt8\") pod \"control-plane-machine-set-operator-78cbb6b69f-6r9rt\" (UID: \"8105a388-83c1-4372-bd51-fcbc2a2e9198\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698541 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698561 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698596 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698616 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpjmk\" (UniqueName: \"kubernetes.io/projected/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-kube-api-access-hpjmk\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698635 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-service-ca\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g7nd\" (UniqueName: \"kubernetes.io/projected/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-kube-api-access-7g7nd\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698673 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-client\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698703 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-config\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698725 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjghs\" (UniqueName: \"kubernetes.io/projected/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-kube-api-access-mjghs\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698755 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-config\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698787 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs25h\" (UniqueName: \"kubernetes.io/projected/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-kube-api-access-gs25h\") pod \"ingress-canary-wj8q6\" (UID: \"146bb9f2-5eb7-43db-a96f-6588ac7bbd12\") " pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698807 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-cert\") pod \"ingress-canary-wj8q6\" (UID: \"146bb9f2-5eb7-43db-a96f-6588ac7bbd12\") " pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698827 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-profile-collector-cert\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698845 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b08695f9-1ea8-4852-8b0a-0d548f000194-metrics-tls\") pod \"dns-operator-744455d44c-v5t7x\" (UID: \"b08695f9-1ea8-4852-8b0a-0d548f000194\") " pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698870 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrsv4\" (UniqueName: \"kubernetes.io/projected/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-kube-api-access-wrsv4\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698888 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-node-bootstrap-token\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698905 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-signing-key\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698931 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c8bd5ed-e597-4095-b5e1-146db751adec-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698952 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-serving-cert\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698969 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.698985 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flsrf\" (UniqueName: \"kubernetes.io/projected/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-kube-api-access-flsrf\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.699007 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.699022 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-certs\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.699037 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-srv-cert\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.699064 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-signing-cabundle\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.699093 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.699109 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-config\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.699125 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8bd5ed-e597-4095-b5e1-146db751adec-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.700588 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-signing-cabundle\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.700667 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.700794 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-config\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.701326 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.702027 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-config\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.702915 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b08695f9-1ea8-4852-8b0a-0d548f000194-metrics-tls\") pod \"dns-operator-744455d44c-v5t7x\" (UID: \"b08695f9-1ea8-4852-8b0a-0d548f000194\") " pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.702917 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.702959 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.703362 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-signing-key\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.703486 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.704401 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.713757 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.736811 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.752566 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.772420 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.781388 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.793179 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.812144 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.829276 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-proxy-tls\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.873034 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.875505 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdq8n\" (UniqueName: \"kubernetes.io/projected/03ba0026-5797-472e-a49c-9789c64cb162-kube-api-access-pdq8n\") pod \"apiserver-7bbb656c7d-9jm8b\" (UID: \"03ba0026-5797-472e-a49c-9789c64cb162\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.892544 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.903310 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8105a388-83c1-4372-bd51-fcbc2a2e9198-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6r9rt\" (UID: \"8105a388-83c1-4372-bd51-fcbc2a2e9198\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.912771 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.932445 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.954663 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.965033 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c8bd5ed-e597-4095-b5e1-146db751adec-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.972431 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.980611 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c8bd5ed-e597-4095-b5e1-146db751adec-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:51 crc kubenswrapper[5035]: I1124 19:16:51.993840 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.013157 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.042675 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.046948 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.052920 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.075319 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.094491 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.103637 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-serving-cert\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.113991 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.126424 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-client\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.133260 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.143443 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-config\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.154916 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.161871 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-ca\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.174438 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.182519 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-etcd-service-ca\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.194565 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.213865 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.232556 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.253235 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.280522 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.292654 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.312714 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.332808 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.353155 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.366562 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-srv-cert\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.372963 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.384284 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.384882 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-profile-collector-cert\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.393109 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.414251 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.433004 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.452621 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.454283 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b"] Nov 24 19:16:52 crc kubenswrapper[5035]: W1124 19:16:52.462767 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03ba0026_5797_472e_a49c_9789c64cb162.slice/crio-c43a6a3e7e002fdfad68525932d1ae303539a321f9c3c28a15f8e26130719ef7 WatchSource:0}: Error finding container c43a6a3e7e002fdfad68525932d1ae303539a321f9c3c28a15f8e26130719ef7: Status 404 returned error can't find the container with id c43a6a3e7e002fdfad68525932d1ae303539a321f9c3c28a15f8e26130719ef7 Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.472618 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.493721 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.511098 5035 request.go:700] Waited for 1.003682842s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-sa-dockercfg-5xfcg&limit=500&resourceVersion=0 Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.513652 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.532807 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.553734 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.574034 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.592940 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.613023 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.633310 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.652700 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.673393 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.691999 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.700839 5035 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.700948 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-certs podName:9f27c073-7ed7-45e0-aa21-2a8d294e2050 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:53.200922968 +0000 UTC m=+151.723429245 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-certs") pod "machine-config-server-2p6pp" (UID: "9f27c073-7ed7-45e0-aa21-2a8d294e2050") : failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.700975 5035 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.700994 5035 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.701035 5035 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.701035 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-node-bootstrap-token podName:9f27c073-7ed7-45e0-aa21-2a8d294e2050 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:53.201015301 +0000 UTC m=+151.723521558 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-node-bootstrap-token") pod "machine-config-server-2p6pp" (UID: "9f27c073-7ed7-45e0-aa21-2a8d294e2050") : failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.701073 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-srv-cert podName:fa17b560-2306-4c7e-ae02-02f6f16db399 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:53.201061822 +0000 UTC m=+151.723568079 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-srv-cert") pod "olm-operator-6b444d44fb-k6sqr" (UID: "fa17b560-2306-4c7e-ae02-02f6f16db399") : failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: E1124 19:16:52.701085 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-cert podName:146bb9f2-5eb7-43db-a96f-6588ac7bbd12 nodeName:}" failed. No retries permitted until 2025-11-24 19:16:53.201079432 +0000 UTC m=+151.723585829 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-cert") pod "ingress-canary-wj8q6" (UID: "146bb9f2-5eb7-43db-a96f-6588ac7bbd12") : failed to sync secret cache: timed out waiting for the condition Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.712431 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.732684 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.752777 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.772682 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.792502 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.812471 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.833016 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.853287 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.873489 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.894218 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.913642 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.933565 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.952885 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.973400 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 19:16:52 crc kubenswrapper[5035]: I1124 19:16:52.994407 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.025431 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.034373 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.053925 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.073884 5035 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.094355 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.114082 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.132694 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.153441 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.173886 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.192410 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.212936 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.219422 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-srv-cert\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.219568 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-cert\") pod \"ingress-canary-wj8q6\" (UID: \"146bb9f2-5eb7-43db-a96f-6588ac7bbd12\") " pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.219611 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-node-bootstrap-token\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.219662 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-certs\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.225024 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-cert\") pod \"ingress-canary-wj8q6\" (UID: \"146bb9f2-5eb7-43db-a96f-6588ac7bbd12\") " pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.225320 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fa17b560-2306-4c7e-ae02-02f6f16db399-srv-cert\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.225770 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-node-bootstrap-token\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.229781 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9f27c073-7ed7-45e0-aa21-2a8d294e2050-certs\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.232906 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.253444 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.272265 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.292958 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.313472 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.381866 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5wjh\" (UniqueName: \"kubernetes.io/projected/2a0d45af-2aba-4830-a849-c2933f94fb83-kube-api-access-z5wjh\") pod \"console-f9d7485db-58jzg\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.403220 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-989gg\" (UniqueName: \"kubernetes.io/projected/0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e-kube-api-access-989gg\") pod \"apiserver-76f77b778f-5z8h5\" (UID: \"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e\") " pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.409363 5035 generic.go:334] "Generic (PLEG): container finished" podID="03ba0026-5797-472e-a49c-9789c64cb162" containerID="f470bcb55eec16e5105f49ba1c210536587f4a142d70cda2a4091ecf8b87c72f" exitCode=0 Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.409406 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" event={"ID":"03ba0026-5797-472e-a49c-9789c64cb162","Type":"ContainerDied","Data":"f470bcb55eec16e5105f49ba1c210536587f4a142d70cda2a4091ecf8b87c72f"} Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.409445 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" event={"ID":"03ba0026-5797-472e-a49c-9789c64cb162","Type":"ContainerStarted","Data":"c43a6a3e7e002fdfad68525932d1ae303539a321f9c3c28a15f8e26130719ef7"} Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.423457 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmvfz\" (UniqueName: \"kubernetes.io/projected/1d176339-2edc-4cf3-ad85-f4a0519f41c2-kube-api-access-zmvfz\") pod \"downloads-7954f5f757-v8j4d\" (UID: \"1d176339-2edc-4cf3-ad85-f4a0519f41c2\") " pod="openshift-console/downloads-7954f5f757-v8j4d" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.440052 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x875\" (UniqueName: \"kubernetes.io/projected/236d3c3c-fe73-409f-b39c-7bb3ba97c902-kube-api-access-4x875\") pod \"machine-api-operator-5694c8668f-lwfhz\" (UID: \"236d3c3c-fe73-409f-b39c-7bb3ba97c902\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.462040 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnczm\" (UniqueName: \"kubernetes.io/projected/170f65fd-91b4-46ef-a97e-dfe96117d8b8-kube-api-access-lnczm\") pod \"cluster-samples-operator-665b6dd947-gl4nm\" (UID: \"170f65fd-91b4-46ef-a97e-dfe96117d8b8\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.476614 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p76v\" (UniqueName: \"kubernetes.io/projected/86c7681b-593c-4e81-8df4-0e90bab1facd-kube-api-access-8p76v\") pod \"authentication-operator-69f744f599-qtsbg\" (UID: \"86c7681b-593c-4e81-8df4-0e90bab1facd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.500422 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hw8n\" (UniqueName: \"kubernetes.io/projected/a4cc55bb-0b42-4a4f-bb69-044000b6ba4b-kube-api-access-5hw8n\") pod \"openshift-apiserver-operator-796bbdcf4f-zdl8j\" (UID: \"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.511394 5035 request.go:700] Waited for 1.909124788s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.514766 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.519013 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pknp2\" (UniqueName: \"kubernetes.io/projected/a6bebe11-7f53-4bb7-b7ea-1e552c4d28be-kube-api-access-pknp2\") pod \"console-operator-58897d9998-8b8fg\" (UID: \"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be\") " pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.526894 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.534204 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhn2c\" (UniqueName: \"kubernetes.io/projected/fb6648af-1ab7-419e-bdb3-8f696aaabd80-kube-api-access-qhn2c\") pod \"oauth-openshift-558db77b4-qfxzs\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.553751 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf257\" (UniqueName: \"kubernetes.io/projected/bb15b826-35be-4433-9e0f-88e013e8141d-kube-api-access-xf257\") pod \"openshift-config-operator-7777fb866f-6b8k5\" (UID: \"bb15b826-35be-4433-9e0f-88e013e8141d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.570418 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc4c4\" (UniqueName: \"kubernetes.io/projected/727a0b65-6c8b-477d-8743-e3bd88a7c904-kube-api-access-kc4c4\") pod \"controller-manager-879f6c89f-r2vkf\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.592554 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.593796 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-868kx\" (UniqueName: \"kubernetes.io/projected/d50060f5-1b00-483a-914b-c34277882c04-kube-api-access-868kx\") pod \"route-controller-manager-6576b87f9c-54m6n\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.611306 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2z6v\" (UniqueName: \"kubernetes.io/projected/43adf3ed-8d96-4dbe-b738-5a7d9fca90f2-kube-api-access-p2z6v\") pod \"machine-approver-56656f9798-sblll\" (UID: \"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.634009 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.634402 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.635770 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c8bd5ed-e597-4095-b5e1-146db751adec-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sv2zq\" (UID: \"3c8bd5ed-e597-4095-b5e1-146db751adec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.646265 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.650837 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flsrf\" (UniqueName: \"kubernetes.io/projected/a58afa5d-79e2-4191-bf3b-397dbf0ee8e6-kube-api-access-flsrf\") pod \"service-ca-9c57cc56f-fk6tx\" (UID: \"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6\") " pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.656668 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.663868 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-v8j4d" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.670542 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae1f4281-2228-43e5-a3e9-14590cb9b2d7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7mkm9\" (UID: \"ae1f4281-2228-43e5-a3e9-14590cb9b2d7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.677412 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.690876 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dszmc\" (UniqueName: \"kubernetes.io/projected/f1ced4c5-722b-4fb9-93fb-37d32b359cf6-kube-api-access-dszmc\") pod \"etcd-operator-b45778765-zfmld\" (UID: \"f1ced4c5-722b-4fb9-93fb-37d32b359cf6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.696828 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.700665 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.711359 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xbv2\" (UniqueName: \"kubernetes.io/projected/9f27c073-7ed7-45e0-aa21-2a8d294e2050-kube-api-access-6xbv2\") pod \"machine-config-server-2p6pp\" (UID: \"9f27c073-7ed7-45e0-aa21-2a8d294e2050\") " pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.732756 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjlt8\" (UniqueName: \"kubernetes.io/projected/8105a388-83c1-4372-bd51-fcbc2a2e9198-kube-api-access-rjlt8\") pod \"control-plane-machine-set-operator-78cbb6b69f-6r9rt\" (UID: \"8105a388-83c1-4372-bd51-fcbc2a2e9198\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.755000 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.755825 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5z8h5"] Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.755963 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdvwz\" (UniqueName: \"kubernetes.io/projected/b08695f9-1ea8-4852-8b0a-0d548f000194-kube-api-access-kdvwz\") pod \"dns-operator-744455d44c-v5t7x\" (UID: \"b08695f9-1ea8-4852-8b0a-0d548f000194\") " pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.770074 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a1cef2e-ded4-4918-b399-bf2a6e63f4ee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7p62f\" (UID: \"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.773958 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.781067 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.781516 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm"] Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.790465 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.792173 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrsv4\" (UniqueName: \"kubernetes.io/projected/dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c-kube-api-access-wrsv4\") pod \"openshift-controller-manager-operator-756b6f6bc6-2zgwt\" (UID: \"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.801099 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.801396 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.803756 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.811935 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jr6b\" (UniqueName: \"kubernetes.io/projected/fa17b560-2306-4c7e-ae02-02f6f16db399-kube-api-access-6jr6b\") pod \"olm-operator-6b444d44fb-k6sqr\" (UID: \"fa17b560-2306-4c7e-ae02-02f6f16db399\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.841986 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs25h\" (UniqueName: \"kubernetes.io/projected/146bb9f2-5eb7-43db-a96f-6588ac7bbd12-kube-api-access-gs25h\") pod \"ingress-canary-wj8q6\" (UID: \"146bb9f2-5eb7-43db-a96f-6588ac7bbd12\") " pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.851025 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qfxzs"] Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.852361 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpjmk\" (UniqueName: \"kubernetes.io/projected/17e347dc-3f72-4ba5-ab67-3e78c4d93eef-kube-api-access-hpjmk\") pod \"machine-config-controller-84d6567774-q8zqf\" (UID: \"17e347dc-3f72-4ba5-ab67-3e78c4d93eef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.872396 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g7nd\" (UniqueName: \"kubernetes.io/projected/eb1f3749-2ef0-4c45-b215-b98ccda4e4a8-kube-api-access-7g7nd\") pod \"kube-storage-version-migrator-operator-b67b599dd-74lhf\" (UID: \"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:53 crc kubenswrapper[5035]: W1124 19:16:53.884424 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43adf3ed_8d96_4dbe_b738_5a7d9fca90f2.slice/crio-8ea40467a951a3dbe8a41b3b780d590e583c14e67b423fb7fabc90f95f1d233e WatchSource:0}: Error finding container 8ea40467a951a3dbe8a41b3b780d590e583c14e67b423fb7fabc90f95f1d233e: Status 404 returned error can't find the container with id 8ea40467a951a3dbe8a41b3b780d590e583c14e67b423fb7fabc90f95f1d233e Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.893109 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjghs\" (UniqueName: \"kubernetes.io/projected/dedcaa39-2f77-4754-ae51-69b47eb8eeb1-kube-api-access-mjghs\") pod \"catalog-operator-68c6474976-bvf2t\" (UID: \"dedcaa39-2f77-4754-ae51-69b47eb8eeb1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:53 crc kubenswrapper[5035]: W1124 19:16:53.897185 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb6648af_1ab7_419e_bdb3_8f696aaabd80.slice/crio-c49f83fa2044b3073d693302b85d5bc2a8eb21f5220d778554c70791bee18b06 WatchSource:0}: Error finding container c49f83fa2044b3073d693302b85d5bc2a8eb21f5220d778554c70791bee18b06: Status 404 returned error can't find the container with id c49f83fa2044b3073d693302b85d5bc2a8eb21f5220d778554c70791bee18b06 Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.921785 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931116 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/87a83e4e-6780-4f2a-9d77-bc87e906fd5f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jhwk9\" (UID: \"87a83e4e-6780-4f2a-9d77-bc87e906fd5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931181 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdhjb\" (UniqueName: \"kubernetes.io/projected/facd9ded-06ed-439c-85ee-b5edd975c1f2-kube-api-access-gdhjb\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931247 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-secret-volume\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931312 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c8636a-40d2-4609-bfb9-808a76cb0f19-service-ca-bundle\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931364 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15eb6fde-d4d1-4ea4-b70f-f956105126a7-trusted-ca\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931419 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-serving-cert\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931444 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vdwm\" (UniqueName: \"kubernetes.io/projected/fa706698-c956-4fe2-aaad-e4feb290d36e-kube-api-access-6vdwm\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931504 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-trusted-ca\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931548 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/facd9ded-06ed-439c-85ee-b5edd975c1f2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931574 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931598 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-config\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931637 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0cb2369e-7c87-4ec1-9b20-f265f921ded5-tmpfs\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931660 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931684 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdpgj\" (UniqueName: \"kubernetes.io/projected/6c737878-3aa7-4291-94dc-55a537dd0fae-kube-api-access-rdpgj\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931739 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-certificates\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931796 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8qrm\" (UniqueName: \"kubernetes.io/projected/87a83e4e-6780-4f2a-9d77-bc87e906fd5f-kube-api-access-t8qrm\") pod \"package-server-manager-789f6589d5-jhwk9\" (UID: \"87a83e4e-6780-4f2a-9d77-bc87e906fd5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931836 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-registration-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931899 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-metrics-tls\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931926 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15eb6fde-d4d1-4ea4-b70f-f956105126a7-metrics-tls\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.931974 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpbqk\" (UniqueName: \"kubernetes.io/projected/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-kube-api-access-cpbqk\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932040 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/facd9ded-06ed-439c-85ee-b5edd975c1f2-images\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932066 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9wm7\" (UniqueName: \"kubernetes.io/projected/3e1e4426-62ad-4454-bbc3-6e117657e33e-kube-api-access-c9wm7\") pod \"multus-admission-controller-857f4d67dd-2j9k6\" (UID: \"3e1e4426-62ad-4454-bbc3-6e117657e33e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932112 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/807e010a-4744-4b1d-aa7c-9cd3360e19af-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932138 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/facd9ded-06ed-439c-85ee-b5edd975c1f2-proxy-tls\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932161 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/15eb6fde-d4d1-4ea4-b70f-f956105126a7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932207 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-plugins-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932733 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8q5g\" (UniqueName: \"kubernetes.io/projected/9f326606-3fdd-4b38-bff6-2f4590b43dcb-kube-api-access-j8q5g\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932768 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-tls\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.932814 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/807e010a-4744-4b1d-aa7c-9cd3360e19af-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.933167 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.933197 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-metrics-certs\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.933291 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f326606-3fdd-4b38-bff6-2f4590b43dcb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.933334 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-config-volume\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:53 crc kubenswrapper[5035]: E1124 19:16:53.933691 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.433678755 +0000 UTC m=+152.956185012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.933709 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj8hp\" (UniqueName: \"kubernetes.io/projected/15eb6fde-d4d1-4ea4-b70f-f956105126a7-kube-api-access-cj8hp\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934067 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z96v\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-kube-api-access-6z96v\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934085 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-default-certificate\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934118 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-bound-sa-token\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934146 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz6hm\" (UniqueName: \"kubernetes.io/projected/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-kube-api-access-kz6hm\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934163 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnxcq\" (UniqueName: \"kubernetes.io/projected/f37e45f1-e73d-4bc0-904c-8ef5354ead61-kube-api-access-cnxcq\") pod \"migrator-59844c95c7-24hv8\" (UID: \"f37e45f1-e73d-4bc0-904c-8ef5354ead61\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934574 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-config-volume\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934599 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f326606-3fdd-4b38-bff6-2f4590b43dcb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934641 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-mountpoint-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-csi-data-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.934720 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-socket-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935108 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cb2369e-7c87-4ec1-9b20-f265f921ded5-webhook-cert\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935152 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm7sl\" (UniqueName: \"kubernetes.io/projected/0cb2369e-7c87-4ec1-9b20-f265f921ded5-kube-api-access-sm7sl\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935241 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e1e4426-62ad-4454-bbc3-6e117657e33e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2j9k6\" (UID: \"3e1e4426-62ad-4454-bbc3-6e117657e33e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935549 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f326606-3fdd-4b38-bff6-2f4590b43dcb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935575 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cb2369e-7c87-4ec1-9b20-f265f921ded5-apiservice-cert\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935761 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjlz\" (UniqueName: \"kubernetes.io/projected/03c8636a-40d2-4609-bfb9-808a76cb0f19-kube-api-access-fnjlz\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935809 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-596xx\" (UniqueName: \"kubernetes.io/projected/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-kube-api-access-596xx\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.935922 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-stats-auth\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.974380 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2p6pp" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.978531 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-wj8q6" Nov 24 19:16:53 crc kubenswrapper[5035]: I1124 19:16:53.985607 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036631 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036743 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z96v\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-kube-api-access-6z96v\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036768 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-default-certificate\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.036811 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.536778763 +0000 UTC m=+153.059285020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036857 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-bound-sa-token\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036900 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz6hm\" (UniqueName: \"kubernetes.io/projected/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-kube-api-access-kz6hm\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnxcq\" (UniqueName: \"kubernetes.io/projected/f37e45f1-e73d-4bc0-904c-8ef5354ead61-kube-api-access-cnxcq\") pod \"migrator-59844c95c7-24hv8\" (UID: \"f37e45f1-e73d-4bc0-904c-8ef5354ead61\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036952 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-config-volume\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.036975 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f326606-3fdd-4b38-bff6-2f4590b43dcb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037068 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-mountpoint-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037124 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-csi-data-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037168 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-socket-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037228 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cb2369e-7c87-4ec1-9b20-f265f921ded5-webhook-cert\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037255 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm7sl\" (UniqueName: \"kubernetes.io/projected/0cb2369e-7c87-4ec1-9b20-f265f921ded5-kube-api-access-sm7sl\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037334 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e1e4426-62ad-4454-bbc3-6e117657e33e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2j9k6\" (UID: \"3e1e4426-62ad-4454-bbc3-6e117657e33e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037365 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f326606-3fdd-4b38-bff6-2f4590b43dcb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037392 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cb2369e-7c87-4ec1-9b20-f265f921ded5-apiservice-cert\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037425 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjlz\" (UniqueName: \"kubernetes.io/projected/03c8636a-40d2-4609-bfb9-808a76cb0f19-kube-api-access-fnjlz\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037452 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-596xx\" (UniqueName: \"kubernetes.io/projected/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-kube-api-access-596xx\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037495 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-stats-auth\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037538 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/87a83e4e-6780-4f2a-9d77-bc87e906fd5f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jhwk9\" (UID: \"87a83e4e-6780-4f2a-9d77-bc87e906fd5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037564 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdhjb\" (UniqueName: \"kubernetes.io/projected/facd9ded-06ed-439c-85ee-b5edd975c1f2-kube-api-access-gdhjb\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037608 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-secret-volume\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037645 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-config-volume\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037679 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-socket-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037716 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c8636a-40d2-4609-bfb9-808a76cb0f19-service-ca-bundle\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037753 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15eb6fde-d4d1-4ea4-b70f-f956105126a7-trusted-ca\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037763 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-mountpoint-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037777 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-serving-cert\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037803 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vdwm\" (UniqueName: \"kubernetes.io/projected/fa706698-c956-4fe2-aaad-e4feb290d36e-kube-api-access-6vdwm\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037848 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-csi-data-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037863 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-trusted-ca\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037890 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/facd9ded-06ed-439c-85ee-b5edd975c1f2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037931 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.037989 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-config\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.038013 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0cb2369e-7c87-4ec1-9b20-f265f921ded5-tmpfs\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.038039 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.038064 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdpgj\" (UniqueName: \"kubernetes.io/projected/6c737878-3aa7-4291-94dc-55a537dd0fae-kube-api-access-rdpgj\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.038089 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-certificates\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.039806 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8qrm\" (UniqueName: \"kubernetes.io/projected/87a83e4e-6780-4f2a-9d77-bc87e906fd5f-kube-api-access-t8qrm\") pod \"package-server-manager-789f6589d5-jhwk9\" (UID: \"87a83e4e-6780-4f2a-9d77-bc87e906fd5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.039892 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-registration-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.039929 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-metrics-tls\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.039958 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15eb6fde-d4d1-4ea4-b70f-f956105126a7-metrics-tls\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.039990 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpbqk\" (UniqueName: \"kubernetes.io/projected/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-kube-api-access-cpbqk\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040054 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9wm7\" (UniqueName: \"kubernetes.io/projected/3e1e4426-62ad-4454-bbc3-6e117657e33e-kube-api-access-c9wm7\") pod \"multus-admission-controller-857f4d67dd-2j9k6\" (UID: \"3e1e4426-62ad-4454-bbc3-6e117657e33e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040087 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/facd9ded-06ed-439c-85ee-b5edd975c1f2-images\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040112 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/807e010a-4744-4b1d-aa7c-9cd3360e19af-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040134 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/facd9ded-06ed-439c-85ee-b5edd975c1f2-proxy-tls\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040157 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/15eb6fde-d4d1-4ea4-b70f-f956105126a7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040178 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-plugins-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040267 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8q5g\" (UniqueName: \"kubernetes.io/projected/9f326606-3fdd-4b38-bff6-2f4590b43dcb-kube-api-access-j8q5g\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040342 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/facd9ded-06ed-439c-85ee-b5edd975c1f2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040396 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-tls\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040433 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/807e010a-4744-4b1d-aa7c-9cd3360e19af-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040464 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040494 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-metrics-certs\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040526 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f326606-3fdd-4b38-bff6-2f4590b43dcb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040553 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-config-volume\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040633 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj8hp\" (UniqueName: \"kubernetes.io/projected/15eb6fde-d4d1-4ea4-b70f-f956105126a7-kube-api-access-cj8hp\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.040712 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c8636a-40d2-4609-bfb9-808a76cb0f19-service-ca-bundle\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.046550 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-plugins-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.046939 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-trusted-ca\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.047840 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f326606-3fdd-4b38-bff6-2f4590b43dcb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.048253 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6c737878-3aa7-4291-94dc-55a537dd0fae-registration-dir\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.051818 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/facd9ded-06ed-439c-85ee-b5edd975c1f2-images\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.051975 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-serving-cert\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.052521 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f326606-3fdd-4b38-bff6-2f4590b43dcb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.053021 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cb2369e-7c87-4ec1-9b20-f265f921ded5-webhook-cert\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.053653 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-config\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.053755 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/87a83e4e-6780-4f2a-9d77-bc87e906fd5f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jhwk9\" (UID: \"87a83e4e-6780-4f2a-9d77-bc87e906fd5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.054517 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/807e010a-4744-4b1d-aa7c-9cd3360e19af-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.056484 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-default-certificate\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.056602 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-secret-volume\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.056921 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0cb2369e-7c87-4ec1-9b20-f265f921ded5-tmpfs\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.057015 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-config-volume\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.057410 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.057676 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.557642707 +0000 UTC m=+153.080148964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.058299 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-certificates\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.058561 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.060715 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15eb6fde-d4d1-4ea4-b70f-f956105126a7-metrics-tls\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.059661 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-tls\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.059786 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.058930 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15eb6fde-d4d1-4ea4-b70f-f956105126a7-trusted-ca\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.064757 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-stats-auth\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.064844 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-metrics-tls\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.065097 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.067462 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/807e010a-4744-4b1d-aa7c-9cd3360e19af-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.067507 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cb2369e-7c87-4ec1-9b20-f265f921ded5-apiservice-cert\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.068797 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03c8636a-40d2-4609-bfb9-808a76cb0f19-metrics-certs\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.069764 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.070219 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3e1e4426-62ad-4454-bbc3-6e117657e33e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2j9k6\" (UID: \"3e1e4426-62ad-4454-bbc3-6e117657e33e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.071674 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z96v\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-kube-api-access-6z96v\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.072834 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.073215 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/facd9ded-06ed-439c-85ee-b5edd975c1f2-proxy-tls\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.090377 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz6hm\" (UniqueName: \"kubernetes.io/projected/82de57d0-1931-4f1a-9ae0-d16aa81c7fd0-kube-api-access-kz6hm\") pod \"dns-default-csmtf\" (UID: \"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0\") " pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.111727 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnxcq\" (UniqueName: \"kubernetes.io/projected/f37e45f1-e73d-4bc0-904c-8ef5354ead61-kube-api-access-cnxcq\") pod \"migrator-59844c95c7-24hv8\" (UID: \"f37e45f1-e73d-4bc0-904c-8ef5354ead61\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.120645 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qtsbg"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.134855 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.139875 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-bound-sa-token\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.140201 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.168592 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.183229 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.669145283 +0000 UTC m=+153.191651530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.190055 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-596xx\" (UniqueName: \"kubernetes.io/projected/5ab1aec8-3cd4-4cea-ab3e-1c710acd1166-kube-api-access-596xx\") pod \"service-ca-operator-777779d784-n5g7f\" (UID: \"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.202015 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vdwm\" (UniqueName: \"kubernetes.io/projected/fa706698-c956-4fe2-aaad-e4feb290d36e-kube-api-access-6vdwm\") pod \"marketplace-operator-79b997595-q2jhw\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.215056 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpbqk\" (UniqueName: \"kubernetes.io/projected/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-kube-api-access-cpbqk\") pod \"collect-profiles-29400195-fd2dn\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:54 crc kubenswrapper[5035]: W1124 19:16:54.242265 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86c7681b_593c_4e81_8df4_0e90bab1facd.slice/crio-fe3140f664a95ec426127837f8de93886f1d94a87bdb10454d3d6c666517d752 WatchSource:0}: Error finding container fe3140f664a95ec426127837f8de93886f1d94a87bdb10454d3d6c666517d752: Status 404 returned error can't find the container with id fe3140f664a95ec426127837f8de93886f1d94a87bdb10454d3d6c666517d752 Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.245599 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9wm7\" (UniqueName: \"kubernetes.io/projected/3e1e4426-62ad-4454-bbc3-6e117657e33e-kube-api-access-c9wm7\") pod \"multus-admission-controller-857f4d67dd-2j9k6\" (UID: \"3e1e4426-62ad-4454-bbc3-6e117657e33e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.251366 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdhjb\" (UniqueName: \"kubernetes.io/projected/facd9ded-06ed-439c-85ee-b5edd975c1f2-kube-api-access-gdhjb\") pod \"machine-config-operator-74547568cd-g6ldj\" (UID: \"facd9ded-06ed-439c-85ee-b5edd975c1f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.274272 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.274606 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.774593504 +0000 UTC m=+153.297099761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.279402 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f326606-3fdd-4b38-bff6-2f4590b43dcb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.284944 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.295416 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm7sl\" (UniqueName: \"kubernetes.io/projected/0cb2369e-7c87-4ec1-9b20-f265f921ded5-kube-api-access-sm7sl\") pod \"packageserver-d55dfcdfc-k7v8b\" (UID: \"0cb2369e-7c87-4ec1-9b20-f265f921ded5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.325002 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8q5g\" (UniqueName: \"kubernetes.io/projected/9f326606-3fdd-4b38-bff6-2f4590b43dcb-kube-api-access-j8q5g\") pod \"cluster-image-registry-operator-dc59b4c8b-k4n59\" (UID: \"9f326606-3fdd-4b38-bff6-2f4590b43dcb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.329682 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdpgj\" (UniqueName: \"kubernetes.io/projected/6c737878-3aa7-4291-94dc-55a537dd0fae-kube-api-access-rdpgj\") pod \"csi-hostpathplugin-vzvml\" (UID: \"6c737878-3aa7-4291-94dc-55a537dd0fae\") " pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.361989 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjlz\" (UniqueName: \"kubernetes.io/projected/03c8636a-40d2-4609-bfb9-808a76cb0f19-kube-api-access-fnjlz\") pod \"router-default-5444994796-w8dwg\" (UID: \"03c8636a-40d2-4609-bfb9-808a76cb0f19\") " pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.371033 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8qrm\" (UniqueName: \"kubernetes.io/projected/87a83e4e-6780-4f2a-9d77-bc87e906fd5f-kube-api-access-t8qrm\") pod \"package-server-manager-789f6589d5-jhwk9\" (UID: \"87a83e4e-6780-4f2a-9d77-bc87e906fd5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.375771 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.375913 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.875892493 +0000 UTC m=+153.398398750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.376194 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.376529 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.876521321 +0000 UTC m=+153.399027578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.390034 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/15eb6fde-d4d1-4ea4-b70f-f956105126a7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.396275 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.412018 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj8hp\" (UniqueName: \"kubernetes.io/projected/15eb6fde-d4d1-4ea4-b70f-f956105126a7-kube-api-access-cj8hp\") pod \"ingress-operator-5b745b69d9-47f49\" (UID: \"15eb6fde-d4d1-4ea4-b70f-f956105126a7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.413501 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.421570 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" event={"ID":"86c7681b-593c-4e81-8df4-0e90bab1facd","Type":"ContainerStarted","Data":"fe3140f664a95ec426127837f8de93886f1d94a87bdb10454d3d6c666517d752"} Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.429050 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.431673 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" event={"ID":"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2","Type":"ContainerStarted","Data":"8ea40467a951a3dbe8a41b3b780d590e583c14e67b423fb7fabc90f95f1d233e"} Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.438156 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2p6pp" event={"ID":"9f27c073-7ed7-45e0-aa21-2a8d294e2050","Type":"ContainerStarted","Data":"bb56fb25dd505c6f6cb714e3d398fe23ca7cad5be71614c2e21763b0d5fd84ae"} Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.449241 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.450909 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" event={"ID":"03ba0026-5797-472e-a49c-9789c64cb162","Type":"ContainerStarted","Data":"955f9a7dc47cbcbcc303f646653d0fb670fb1608fec5201a8e7da9493090a988"} Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.455254 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" event={"ID":"fb6648af-1ab7-419e-bdb3-8f696aaabd80","Type":"ContainerStarted","Data":"c49f83fa2044b3073d693302b85d5bc2a8eb21f5220d778554c70791bee18b06"} Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.463768 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" event={"ID":"170f65fd-91b4-46ef-a97e-dfe96117d8b8","Type":"ContainerStarted","Data":"581dea44a352558eb694ba4899ee5f44ab7dc9ff1c6901890ba56d385828f6f2"} Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.465777 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.473709 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.477275 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.477718 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:54.977697216 +0000 UTC m=+153.500203473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.480404 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" event={"ID":"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e","Type":"ContainerStarted","Data":"cd30a39139598a61d864a4d377045d8560790d29a6894065396fd23a02508a07"} Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.483666 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.495701 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.506719 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.513145 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.520990 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.521039 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-58jzg"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.538816 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.547981 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.570465 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-v8j4d"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.577333 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8b8fg"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.580924 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.581532 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.081514134 +0000 UTC m=+153.604020391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.682115 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.682634 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.182613488 +0000 UTC m=+153.705119745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.699467 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-lwfhz"] Nov 24 19:16:54 crc kubenswrapper[5035]: W1124 19:16:54.716879 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6bebe11_7f53_4bb7_b7ea_1e552c4d28be.slice/crio-39d2cd299fe0de34550331ceb528dc96dd1573e5b1d3f574b3f9023031dd3eb3 WatchSource:0}: Error finding container 39d2cd299fe0de34550331ceb528dc96dd1573e5b1d3f574b3f9023031dd3eb3: Status 404 returned error can't find the container with id 39d2cd299fe0de34550331ceb528dc96dd1573e5b1d3f574b3f9023031dd3eb3 Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.785040 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.786039 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.286005034 +0000 UTC m=+153.808511291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: W1124 19:16:54.823068 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod236d3c3c_fe73_409f_b39c_7bb3ba97c902.slice/crio-eed7709f976a311cf902751a92633f7dd097540defa2ad0ded9a09da6a81331f WatchSource:0}: Error finding container eed7709f976a311cf902751a92633f7dd097540defa2ad0ded9a09da6a81331f: Status 404 returned error can't find the container with id eed7709f976a311cf902751a92633f7dd097540defa2ad0ded9a09da6a81331f Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.845998 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.875331 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.902832 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:54 crc kubenswrapper[5035]: E1124 19:16:54.903336 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.403307456 +0000 UTC m=+153.925813713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.914825 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fk6tx"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.936180 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-zfmld"] Nov 24 19:16:54 crc kubenswrapper[5035]: I1124 19:16:54.957065 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.006701 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.008946 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.508930753 +0000 UTC m=+154.031437010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.028028 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.032224 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.053619 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-wj8q6"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.069004 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.108150 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.108378 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.608354612 +0000 UTC m=+154.130860869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.108704 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.109738 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.609005399 +0000 UTC m=+154.131511656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: W1124 19:16:55.141174 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c8bd5ed_e597_4095_b5e1_146db751adec.slice/crio-b8de5ad0894da841556fe4445235d8a7c26641cf1adb1f8920c9ffa039f0410c WatchSource:0}: Error finding container b8de5ad0894da841556fe4445235d8a7c26641cf1adb1f8920c9ffa039f0410c: Status 404 returned error can't find the container with id b8de5ad0894da841556fe4445235d8a7c26641cf1adb1f8920c9ffa039f0410c Nov 24 19:16:55 crc kubenswrapper[5035]: W1124 19:16:55.182242 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda58afa5d_79e2_4191_bf3b_397dbf0ee8e6.slice/crio-b948df05a80c0fcb5bd299d31934db5c952de6f7b5f90a456911b67aaf5ea8cd WatchSource:0}: Error finding container b948df05a80c0fcb5bd299d31934db5c952de6f7b5f90a456911b67aaf5ea8cd: Status 404 returned error can't find the container with id b948df05a80c0fcb5bd299d31934db5c952de6f7b5f90a456911b67aaf5ea8cd Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.209767 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.210079 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.710061961 +0000 UTC m=+154.232568208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.311075 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.311481 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.811466774 +0000 UTC m=+154.333973021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.412561 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.413115 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:55.913078471 +0000 UTC m=+154.435584728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.513713 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.513987 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.01397678 +0000 UTC m=+154.536483037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.522530 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.558425 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.618130 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.619057 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.119041101 +0000 UTC m=+154.641547358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.621810 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.632166 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-47f49"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.659201 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.663230 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v5t7x"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.664119 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" event={"ID":"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee","Type":"ContainerStarted","Data":"67411ca21c33277e6b47eef6a1f6e93c91497e438e7bae4548227f6fbb225c29"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.668657 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r2vkf"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.673226 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" event={"ID":"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6","Type":"ContainerStarted","Data":"b948df05a80c0fcb5bd299d31934db5c952de6f7b5f90a456911b67aaf5ea8cd"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.676189 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.679498 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" event={"ID":"d50060f5-1b00-483a-914b-c34277882c04","Type":"ContainerStarted","Data":"759e37b9e45bcc763b50d722427ac4522c82991da3c178214e3e70036545044f"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.683490 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-v8j4d" event={"ID":"1d176339-2edc-4cf3-ad85-f4a0519f41c2","Type":"ContainerStarted","Data":"47d2ad949a46a5d1d797bfbe8f4cde23f739a70447344329f3593f13fc6851de"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.684793 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-v8j4d" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.685822 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-v8j4d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.685860 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-v8j4d" podUID="1d176339-2edc-4cf3-ad85-f4a0519f41c2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.688610 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" event={"ID":"236d3c3c-fe73-409f-b39c-7bb3ba97c902","Type":"ContainerStarted","Data":"eed7709f976a311cf902751a92633f7dd097540defa2ad0ded9a09da6a81331f"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.691472 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" event={"ID":"fb6648af-1ab7-419e-bdb3-8f696aaabd80","Type":"ContainerStarted","Data":"6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.692883 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.693569 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" event={"ID":"ae1f4281-2228-43e5-a3e9-14590cb9b2d7","Type":"ContainerStarted","Data":"9c6a7d8fb617045a23a35839393c9b72a73a59289c724ef6767241e441afbe31"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.697291 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.698224 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" event={"ID":"bb15b826-35be-4433-9e0f-88e013e8141d","Type":"ContainerStarted","Data":"2da606d9cf5da720cd6eda5468ad4e1622b485ec4041325d861b9ee0f145f043"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.707255 5035 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-qfxzs container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.707324 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" podUID="fb6648af-1ab7-419e-bdb3-8f696aaabd80" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.708820 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58jzg" event={"ID":"2a0d45af-2aba-4830-a849-c2933f94fb83","Type":"ContainerStarted","Data":"71e590e92e30f69e7d0c51233569a82504b33552e6054792779a19474ef433cb"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.710226 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.712969 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" event={"ID":"86c7681b-593c-4e81-8df4-0e90bab1facd","Type":"ContainerStarted","Data":"a2464e9b0e456652f67cde1350a095f213abdfe196764cf65b6bfe66413698b7"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.713040 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.717987 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" event={"ID":"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be","Type":"ContainerStarted","Data":"39d2cd299fe0de34550331ceb528dc96dd1573e5b1d3f574b3f9023031dd3eb3"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.718121 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.719966 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.720275 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.220263259 +0000 UTC m=+154.742769516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.720386 5035 patch_prober.go:28] interesting pod/console-operator-58897d9998-8b8fg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.720408 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" podUID="a6bebe11-7f53-4bb7-b7ea-1e552c4d28be" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.722301 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.730349 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.734834 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" event={"ID":"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2","Type":"ContainerStarted","Data":"6358ab8c7d9909fe39e65207cb40e5743f1d70c0eacba28260fc8100395fc3a6"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.736362 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.741687 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" event={"ID":"170f65fd-91b4-46ef-a97e-dfe96117d8b8","Type":"ContainerStarted","Data":"6e08ac4849c011978b73295d47d9beddb0a847ef8f6369769b9ca3994f6d4ddb"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.741734 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" event={"ID":"170f65fd-91b4-46ef-a97e-dfe96117d8b8","Type":"ContainerStarted","Data":"2f142085aab696467fb9f92bd89e634a1496b5cf801ca404cd4300b17dc84269"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.749192 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q2jhw"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.751187 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" event={"ID":"fa17b560-2306-4c7e-ae02-02f6f16db399","Type":"ContainerStarted","Data":"856c92fd6adce6b4dbd612988810130665ed2595c7ecafbbbdc3d8ecead53dd4"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.752010 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.754663 5035 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k6sqr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.754706 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" podUID="fa17b560-2306-4c7e-ae02-02f6f16db399" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.758206 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-csmtf"] Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.758929 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" event={"ID":"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b","Type":"ContainerStarted","Data":"a2c23747fbdc89646041c176941503a9f66ec6436c6b88f59cec60cd14ed8410"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.762172 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w8dwg" event={"ID":"03c8636a-40d2-4609-bfb9-808a76cb0f19","Type":"ContainerStarted","Data":"b0436bd8ce1b696411a61af2d6e63e32fccbc62bf0f3ea87c5d4af950a3beb5d"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.767499 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2p6pp" event={"ID":"9f27c073-7ed7-45e0-aa21-2a8d294e2050","Type":"ContainerStarted","Data":"68e8f50b2b0ea822a4d205452055ac33559715104410ee8fe0d56d4e35e5b7ab"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.769717 5035 generic.go:334] "Generic (PLEG): container finished" podID="0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e" containerID="4d416d35ed979cf3e2db2f686829ba51a282086c1ffe6cae0c25eda8a7bd03bc" exitCode=0 Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.769774 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" event={"ID":"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e","Type":"ContainerDied","Data":"4d416d35ed979cf3e2db2f686829ba51a282086c1ffe6cae0c25eda8a7bd03bc"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.771867 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" event={"ID":"f1ced4c5-722b-4fb9-93fb-37d32b359cf6","Type":"ContainerStarted","Data":"0baafa11558ffb7324e8de647adeb3ad576e1fb4e235c7f33227850de879e7f3"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.777574 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" event={"ID":"3c8bd5ed-e597-4095-b5e1-146db751adec","Type":"ContainerStarted","Data":"b8de5ad0894da841556fe4445235d8a7c26641cf1adb1f8920c9ffa039f0410c"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.780262 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" event={"ID":"8105a388-83c1-4372-bd51-fcbc2a2e9198","Type":"ContainerStarted","Data":"8124f58a7aa35fbd74b377658c3b81249f964476c5b52a4dc14b461b32852c64"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.792926 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-wj8q6" event={"ID":"146bb9f2-5eb7-43db-a96f-6588ac7bbd12","Type":"ContainerStarted","Data":"d0c656a7e8f4ba78f922a9117fe01f1e654582e5cae3c9d6435784c7eb20d271"} Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.811398 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vzvml"] Nov 24 19:16:55 crc kubenswrapper[5035]: W1124 19:16:55.811825 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddadbc1b7_3a9b_4be3_b7c1_c86fc67e6d1c.slice/crio-c11f1597682f3a31da42a2eef67654c59f6b1492584992c00ae98889ddda09f6 WatchSource:0}: Error finding container c11f1597682f3a31da42a2eef67654c59f6b1492584992c00ae98889ddda09f6: Status 404 returned error can't find the container with id c11f1597682f3a31da42a2eef67654c59f6b1492584992c00ae98889ddda09f6 Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.820586 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.820723 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.320702864 +0000 UTC m=+154.843209121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.821261 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.826308 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.326285185 +0000 UTC m=+154.848791442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: W1124 19:16:55.843819 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa706698_c956_4fe2_aaad_e4feb290d36e.slice/crio-40208ea8593284174e68f425948d633557e6c419e14f135ea7f9673c329deeee WatchSource:0}: Error finding container 40208ea8593284174e68f425948d633557e6c419e14f135ea7f9673c329deeee: Status 404 returned error can't find the container with id 40208ea8593284174e68f425948d633557e6c419e14f135ea7f9673c329deeee Nov 24 19:16:55 crc kubenswrapper[5035]: W1124 19:16:55.850583 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87a83e4e_6780_4f2a_9d77_bc87e906fd5f.slice/crio-baf665aaaa23bd2f22ec21642fc77b472c3f3451242616c756a8907f0f74074e WatchSource:0}: Error finding container baf665aaaa23bd2f22ec21642fc77b472c3f3451242616c756a8907f0f74074e: Status 404 returned error can't find the container with id baf665aaaa23bd2f22ec21642fc77b472c3f3451242616c756a8907f0f74074e Nov 24 19:16:55 crc kubenswrapper[5035]: W1124 19:16:55.852120 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15eb6fde_d4d1_4ea4_b70f_f956105126a7.slice/crio-9016a57fc31cb885a33a01cc4a9d92f7023100eee2f968e8b1bb417cc6b1c375 WatchSource:0}: Error finding container 9016a57fc31cb885a33a01cc4a9d92f7023100eee2f968e8b1bb417cc6b1c375: Status 404 returned error can't find the container with id 9016a57fc31cb885a33a01cc4a9d92f7023100eee2f968e8b1bb417cc6b1c375 Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.855706 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2j9k6"] Nov 24 19:16:55 crc kubenswrapper[5035]: W1124 19:16:55.864460 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddedcaa39_2f77_4754_ae51_69b47eb8eeb1.slice/crio-899006a30bad763ccbd1d36efe2443d66d0e4dc58e17e564b58d08e4f05b267b WatchSource:0}: Error finding container 899006a30bad763ccbd1d36efe2443d66d0e4dc58e17e564b58d08e4f05b267b: Status 404 returned error can't find the container with id 899006a30bad763ccbd1d36efe2443d66d0e4dc58e17e564b58d08e4f05b267b Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.869810 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" podStartSLOduration=128.869789472 podStartE2EDuration="2m8.869789472s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:55.867797108 +0000 UTC m=+154.390303385" watchObservedRunningTime="2025-11-24 19:16:55.869789472 +0000 UTC m=+154.392295739" Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.924952 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.944642 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.444581305 +0000 UTC m=+154.967087592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:55 crc kubenswrapper[5035]: I1124 19:16:55.945507 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:55 crc kubenswrapper[5035]: E1124 19:16:55.946525 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.446502796 +0000 UTC m=+154.969009053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.047040 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.047398 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.547376014 +0000 UTC m=+155.069882271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.065061 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gl4nm" podStartSLOduration=130.065047343 podStartE2EDuration="2m10.065047343s" podCreationTimestamp="2025-11-24 19:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.062870273 +0000 UTC m=+154.585376530" watchObservedRunningTime="2025-11-24 19:16:56.065047343 +0000 UTC m=+154.587553600" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.141849 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" podStartSLOduration=130.141820839 podStartE2EDuration="2m10.141820839s" podCreationTimestamp="2025-11-24 19:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.140344869 +0000 UTC m=+154.662851126" watchObservedRunningTime="2025-11-24 19:16:56.141820839 +0000 UTC m=+154.664327096" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.148349 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.148616 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.648605743 +0000 UTC m=+155.171112000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.236535 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-58jzg" podStartSLOduration=129.236520259 podStartE2EDuration="2m9.236520259s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.192390176 +0000 UTC m=+154.714896433" watchObservedRunningTime="2025-11-24 19:16:56.236520259 +0000 UTC m=+154.759026516" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.249759 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.250108 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.750093317 +0000 UTC m=+155.272599574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.350834 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.351201 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.85118785 +0000 UTC m=+155.373694107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.353872 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2p6pp" podStartSLOduration=5.353858143 podStartE2EDuration="5.353858143s" podCreationTimestamp="2025-11-24 19:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.314153639 +0000 UTC m=+154.836659896" watchObservedRunningTime="2025-11-24 19:16:56.353858143 +0000 UTC m=+154.876364400" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.451786 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.452137 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:56.952119539 +0000 UTC m=+155.474625796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.485064 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.489038 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:16:56 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:16:56 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:16:56 crc kubenswrapper[5035]: healthz check failed Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.489115 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.552934 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.553195 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.053184593 +0000 UTC m=+155.575690850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.654175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.664698 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.164661037 +0000 UTC m=+155.687167294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.665082 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.665489 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.16548172 +0000 UTC m=+155.687987977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.712152 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" podStartSLOduration=129.712130791 podStartE2EDuration="2m9.712130791s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.710330283 +0000 UTC m=+155.232836540" watchObservedRunningTime="2025-11-24 19:16:56.712130791 +0000 UTC m=+155.234637048" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.765955 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.766738 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.266717047 +0000 UTC m=+155.789223304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.787741 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" podStartSLOduration=129.787720925 podStartE2EDuration="2m9.787720925s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.787081758 +0000 UTC m=+155.309588025" watchObservedRunningTime="2025-11-24 19:16:56.787720925 +0000 UTC m=+155.310227182" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.837590 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" event={"ID":"87a83e4e-6780-4f2a-9d77-bc87e906fd5f","Type":"ContainerStarted","Data":"bbd512740792249fbc78bf585a27fa608fde1ca11f6914c707c8308de3713f07"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.837643 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" event={"ID":"87a83e4e-6780-4f2a-9d77-bc87e906fd5f","Type":"ContainerStarted","Data":"baf665aaaa23bd2f22ec21642fc77b472c3f3451242616c756a8907f0f74074e"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.846633 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-v8j4d" podStartSLOduration=129.846597727 podStartE2EDuration="2m9.846597727s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.84411181 +0000 UTC m=+155.366618067" watchObservedRunningTime="2025-11-24 19:16:56.846597727 +0000 UTC m=+155.369103984" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.852018 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" event={"ID":"6c737878-3aa7-4291-94dc-55a537dd0fae","Type":"ContainerStarted","Data":"d9e23254d8d98de6677c8312a931a30ab71df2dc0171c103f4167597f4850218"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.869571 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" event={"ID":"a58afa5d-79e2-4191-bf3b-397dbf0ee8e6","Type":"ContainerStarted","Data":"3a283118b0f6196da93b90f2a0cfafd897fa943e45c8d417598b726957b1fd13"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.870955 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qtsbg" podStartSLOduration=130.870938945 podStartE2EDuration="2m10.870938945s" podCreationTimestamp="2025-11-24 19:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:56.870833723 +0000 UTC m=+155.393339970" watchObservedRunningTime="2025-11-24 19:16:56.870938945 +0000 UTC m=+155.393445202" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.871428 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.871719 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.371709197 +0000 UTC m=+155.894215454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.909299 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" event={"ID":"727a0b65-6c8b-477d-8743-e3bd88a7c904","Type":"ContainerStarted","Data":"c801e3af33bd17691ce109e5f20a5e66e798b3ea03fa9d9e08c8abdc1fb0c638"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.909356 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" event={"ID":"727a0b65-6c8b-477d-8743-e3bd88a7c904","Type":"ContainerStarted","Data":"477ba616681c6cefbcae9b560c9fe2fb1a3dbc6e05a341dcfdbb44a1b4d11101"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.909373 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.919205 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" event={"ID":"2a1cef2e-ded4-4918-b399-bf2a6e63f4ee","Type":"ContainerStarted","Data":"cdd01e3389711b8ff230ef0b1e98473a072e0ce72a272a63761b5edcfdc0d710"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.920299 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" event={"ID":"ae1f4281-2228-43e5-a3e9-14590cb9b2d7","Type":"ContainerStarted","Data":"b9931c8f89e5eda0087d6ef4a821a86a645eb4b6936feeb5248acb311eb80ace"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.921626 5035 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-r2vkf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.921679 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" podUID="727a0b65-6c8b-477d-8743-e3bd88a7c904" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.944946 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" event={"ID":"d50060f5-1b00-483a-914b-c34277882c04","Type":"ContainerStarted","Data":"cf65f0579a5e7c5698cf54ba5f7a5f6ed89d1f87a826463853549c1d2a841d4e"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.945761 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.948973 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-csmtf" event={"ID":"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0","Type":"ContainerStarted","Data":"dda52a8e68fadf18a91b55d0007961cb0110817ad87361f92b3ceae14e3b59fd"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.949725 5035 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-54m6n container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.949799 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" podUID="d50060f5-1b00-483a-914b-c34277882c04" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.973590 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:56 crc kubenswrapper[5035]: E1124 19:16:56.974854 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.474832736 +0000 UTC m=+155.997338993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.978376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" event={"ID":"facd9ded-06ed-439c-85ee-b5edd975c1f2","Type":"ContainerStarted","Data":"9bae50d7bc25caf03b560bbf3f3dd68e17dd9e41e33caaa55c47b8ef07410a17"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.978945 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" event={"ID":"facd9ded-06ed-439c-85ee-b5edd975c1f2","Type":"ContainerStarted","Data":"201c143cac39b78cd40482763b642475ea50079a69aff2e888839272afd18b50"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.983111 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" event={"ID":"dedcaa39-2f77-4754-ae51-69b47eb8eeb1","Type":"ContainerStarted","Data":"899006a30bad763ccbd1d36efe2443d66d0e4dc58e17e564b58d08e4f05b267b"} Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.983669 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.984760 5035 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-bvf2t container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Nov 24 19:16:56 crc kubenswrapper[5035]: I1124 19:16:56.984826 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" podUID="dedcaa39-2f77-4754-ae51-69b47eb8eeb1" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.021217 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" event={"ID":"17e347dc-3f72-4ba5-ab67-3e78c4d93eef","Type":"ContainerStarted","Data":"f2c610afa018ca3c2d0b05b83c47cb1d1170964608ad52ee01db3063f008fb79"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.021260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" event={"ID":"17e347dc-3f72-4ba5-ab67-3e78c4d93eef","Type":"ContainerStarted","Data":"4a355fa4b913379ed4a8f7ad598ccca18546593ccb61be7250d3521c6c0381ab"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.021279 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" event={"ID":"17e347dc-3f72-4ba5-ab67-3e78c4d93eef","Type":"ContainerStarted","Data":"27df72cb40629c5f39bca0427a20f0f914db38fc1c16602df9c61e4879f64d97"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.048607 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w8dwg" event={"ID":"03c8636a-40d2-4609-bfb9-808a76cb0f19","Type":"ContainerStarted","Data":"b75f6ecdb0ac41a903d2a7f3698728c8c138925168fb4bd09015796df835a421"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.051713 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.052025 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.071438 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" event={"ID":"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166","Type":"ContainerStarted","Data":"dc70722050af057fb6f09324f1d5030b8b2d770d0c569b485e9b0ba8f7e46a0e"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.071476 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" event={"ID":"5ab1aec8-3cd4-4cea-ab3e-1c710acd1166","Type":"ContainerStarted","Data":"cdb0f1b50585334a00df73b6043b2d45cc956e5d14bfe8a264438535f7231f5c"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.072628 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.076923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.078474 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.578462238 +0000 UTC m=+156.100968495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.085195 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" event={"ID":"0cb2369e-7c87-4ec1-9b20-f265f921ded5","Type":"ContainerStarted","Data":"55009438f24ce5c6c6c2b0952eb34467605743c7ff96e0751e0e8b78b863a982"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.085242 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" event={"ID":"0cb2369e-7c87-4ec1-9b20-f265f921ded5","Type":"ContainerStarted","Data":"29096f793dc13785de3495f002284447e18f6601ca74d3116d783191a6680c0b"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.086031 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.092282 5035 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-k7v8b container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.092351 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" podUID="0cb2369e-7c87-4ec1-9b20-f265f921ded5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.097915 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sblll" event={"ID":"43adf3ed-8d96-4dbe-b738-5a7d9fca90f2","Type":"ContainerStarted","Data":"3a4ff6df9a3abab0a40871b4e8c6089e17286c9cc4f0b46beac9952aa7f44bd0"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.147590 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" event={"ID":"fa17b560-2306-4c7e-ae02-02f6f16db399","Type":"ContainerStarted","Data":"aa00a2f8d8ded9964482b3a71049cc7b9fa992968adc0d2b7af33d57a5b35f89"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.163953 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" event={"ID":"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c","Type":"ContainerStarted","Data":"54c3fbd37eb7794095db427c54339ec1f2abeb34c76eb72364e35e56d0aa7fe3"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.163991 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" event={"ID":"dadbc1b7-3a9b-4be3-b7c1-c86fc67e6d1c","Type":"ContainerStarted","Data":"c11f1597682f3a31da42a2eef67654c59f6b1492584992c00ae98889ddda09f6"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.165082 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" podStartSLOduration=130.16506695 podStartE2EDuration="2m10.16506695s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.12217301 +0000 UTC m=+155.644679267" watchObservedRunningTime="2025-11-24 19:16:57.16506695 +0000 UTC m=+155.687573197" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.170010 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k6sqr" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.178791 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.182966 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.682941713 +0000 UTC m=+156.205447960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.232938 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-w8dwg" podStartSLOduration=130.232922145 podStartE2EDuration="2m10.232922145s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.164279608 +0000 UTC m=+155.686785865" watchObservedRunningTime="2025-11-24 19:16:57.232922145 +0000 UTC m=+155.755428402" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.233689 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" podStartSLOduration=131.233680326 podStartE2EDuration="2m11.233680326s" podCreationTimestamp="2025-11-24 19:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.231983239 +0000 UTC m=+155.754489506" watchObservedRunningTime="2025-11-24 19:16:57.233680326 +0000 UTC m=+155.756186583" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.235516 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" event={"ID":"b08695f9-1ea8-4852-8b0a-0d548f000194","Type":"ContainerStarted","Data":"95132c59d813398915c58a19f33557fbf9561a9e73673377f35f523669501c96"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.235547 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" event={"ID":"b08695f9-1ea8-4852-8b0a-0d548f000194","Type":"ContainerStarted","Data":"8841ff04cde1249f38bf4cf0480d5ada496a17761b4ba4f4ef89f77c0ab80c0b"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.260647 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" event={"ID":"f1ced4c5-722b-4fb9-93fb-37d32b359cf6","Type":"ContainerStarted","Data":"f693d2fbf571032d392deabf7cb6aaf63a0d8c079cf1e1a6f7027f4ec8299e0c"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.261975 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" event={"ID":"9f326606-3fdd-4b38-bff6-2f4590b43dcb","Type":"ContainerStarted","Data":"16e366a2e8ccc8fc378310e45c1fbaa4e838ac3b2dc32cb7d306201cf0aaec68"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.261998 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" event={"ID":"9f326606-3fdd-4b38-bff6-2f4590b43dcb","Type":"ContainerStarted","Data":"9c46180f19a88ea8da930cde55b333665f2509da10604cd2d44f4a04a45f8757"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.264210 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" event={"ID":"3e1e4426-62ad-4454-bbc3-6e117657e33e","Type":"ContainerStarted","Data":"a3fbeb1aee816fb5badf851a4e60b5ff3498c5bb96d83228d9dd37bd7a28c226"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.264261 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" event={"ID":"3e1e4426-62ad-4454-bbc3-6e117657e33e","Type":"ContainerStarted","Data":"34d9de74e54b6127559fda6fd5949f1565b300348110c619de837ea7dac00997"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.269726 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-v8j4d" event={"ID":"1d176339-2edc-4cf3-ad85-f4a0519f41c2","Type":"ContainerStarted","Data":"37b71da86120ad072da14cf71fd2246a57e8b8c130ad4561d9b714bd120f5c2d"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.289193 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.289564 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-v8j4d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.300396 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-v8j4d" podUID="1d176339-2edc-4cf3-ad85-f4a0519f41c2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.295515 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" event={"ID":"a6bebe11-7f53-4bb7-b7ea-1e552c4d28be","Type":"ContainerStarted","Data":"4ceb761240a398a2acc5d91ed9d246bb4a67d7879b2e58c8306288cdf248de67"} Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.290923 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.790910852 +0000 UTC m=+156.313417109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.296449 5035 patch_prober.go:28] interesting pod/console-operator-58897d9998-8b8fg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.300997 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" podUID="a6bebe11-7f53-4bb7-b7ea-1e552c4d28be" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.320433 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" podStartSLOduration=131.32040175 podStartE2EDuration="2m11.32040175s" podCreationTimestamp="2025-11-24 19:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.319469725 +0000 UTC m=+155.841975982" watchObservedRunningTime="2025-11-24 19:16:57.32040175 +0000 UTC m=+155.842908007" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.352401 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" event={"ID":"fa706698-c956-4fe2-aaad-e4feb290d36e","Type":"ContainerStarted","Data":"e116fe9499a7fe2502e95703f1a7252a24d2393944938a45c46e89aa2451ce8d"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.352447 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" event={"ID":"fa706698-c956-4fe2-aaad-e4feb290d36e","Type":"ContainerStarted","Data":"40208ea8593284174e68f425948d633557e6c419e14f135ea7f9673c329deeee"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.353505 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.363876 5035 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q2jhw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.363919 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.397085 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.397412 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6r9rt" event={"ID":"8105a388-83c1-4372-bd51-fcbc2a2e9198","Type":"ContainerStarted","Data":"e3ed8fd8c9fe0705191f53100595aa4b7ea4620b32adc2608af0a703b2286b63"} Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.398218 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:57.898195854 +0000 UTC m=+156.420702111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.411439 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-fk6tx" podStartSLOduration=130.411411681 podStartE2EDuration="2m10.411411681s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.36367076 +0000 UTC m=+155.886177017" watchObservedRunningTime="2025-11-24 19:16:57.411411681 +0000 UTC m=+155.933917958" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.414734 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2zgwt" podStartSLOduration=130.414710681 podStartE2EDuration="2m10.414710681s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.402568372 +0000 UTC m=+155.925074619" watchObservedRunningTime="2025-11-24 19:16:57.414710681 +0000 UTC m=+155.937216928" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.427577 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-wj8q6" event={"ID":"146bb9f2-5eb7-43db-a96f-6588ac7bbd12","Type":"ContainerStarted","Data":"5e8e6c4d7ef44c590c27eee3a094e230722e3a313ced0d16bdc90f782bbe9ce6"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.467056 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58jzg" event={"ID":"2a0d45af-2aba-4830-a849-c2933f94fb83","Type":"ContainerStarted","Data":"0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.484568 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" podStartSLOduration=130.48455352 podStartE2EDuration="2m10.48455352s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.483833 +0000 UTC m=+156.006339257" watchObservedRunningTime="2025-11-24 19:16:57.48455352 +0000 UTC m=+156.007059777" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.499853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.500347 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.000335176 +0000 UTC m=+156.522841433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.506174 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:16:57 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:16:57 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:16:57 crc kubenswrapper[5035]: healthz check failed Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.506231 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.526692 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" event={"ID":"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8","Type":"ContainerStarted","Data":"6407815a4e952812a1df844fe4f91fd52e075a3d1b1e2d9c15a0d89531f0bf11"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.526767 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" event={"ID":"eb1f3749-2ef0-4c45-b215-b98ccda4e4a8","Type":"ContainerStarted","Data":"242e7a2fb57a24d14da26b9df40d3ed8f29775da929d7bcc5eb0029a37ea04ad"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.554885 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zdl8j" event={"ID":"a4cc55bb-0b42-4a4f-bb69-044000b6ba4b","Type":"ContainerStarted","Data":"e52f04b7049adc2fee41fed94649a15ead7e3fa1400a3672d94da1c3eb772a12"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.565501 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q8zqf" podStartSLOduration=130.565485388 podStartE2EDuration="2m10.565485388s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.518778805 +0000 UTC m=+156.041285062" watchObservedRunningTime="2025-11-24 19:16:57.565485388 +0000 UTC m=+156.087991645" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.597509 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4n59" podStartSLOduration=130.597489363 podStartE2EDuration="2m10.597489363s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.595560651 +0000 UTC m=+156.118066908" watchObservedRunningTime="2025-11-24 19:16:57.597489363 +0000 UTC m=+156.119995620" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.602400 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.603817 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.103789534 +0000 UTC m=+156.626295861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.608630 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" event={"ID":"15eb6fde-d4d1-4ea4-b70f-f956105126a7","Type":"ContainerStarted","Data":"2056feca85801057e0d33dda8e85114bc42982c6c3b41078e37b749a2958f88b"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.608676 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" event={"ID":"15eb6fde-d4d1-4ea4-b70f-f956105126a7","Type":"ContainerStarted","Data":"9016a57fc31cb885a33a01cc4a9d92f7023100eee2f968e8b1bb417cc6b1c375"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.630229 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n5g7f" podStartSLOduration=130.630213719 podStartE2EDuration="2m10.630213719s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.626174809 +0000 UTC m=+156.148681066" watchObservedRunningTime="2025-11-24 19:16:57.630213719 +0000 UTC m=+156.152719976" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.630653 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" event={"ID":"f37e45f1-e73d-4bc0-904c-8ef5354ead61","Type":"ContainerStarted","Data":"9b0d15af75f49776e6f5fff1d6b692320dc4cfdaee99969ccbbfd84d88ec2b7f"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.632215 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" event={"ID":"f37e45f1-e73d-4bc0-904c-8ef5354ead61","Type":"ContainerStarted","Data":"21ca7a242c41b4b87e0c3cbbf206a60c7bea3d077387e1c2e68aae68817c82e9"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.661881 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" event={"ID":"236d3c3c-fe73-409f-b39c-7bb3ba97c902","Type":"ContainerStarted","Data":"dbcd76bc2bd20344a3f6ec9ed631fea8e5bc5373aea82e026aa65dbac7a8abb4"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.661932 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" event={"ID":"236d3c3c-fe73-409f-b39c-7bb3ba97c902","Type":"ContainerStarted","Data":"a6c632faa61276a83a60cf3aaf21ab43e2fb88ec52701642f0e17db71f791cf8"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.676934 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" event={"ID":"3c8bd5ed-e597-4095-b5e1-146db751adec","Type":"ContainerStarted","Data":"009bd171a3ca52829c6c42f62da327f96ec47541664b894eae029366579eb348"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.678673 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" podStartSLOduration=130.678655808 podStartE2EDuration="2m10.678655808s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.648817601 +0000 UTC m=+156.171323858" watchObservedRunningTime="2025-11-24 19:16:57.678655808 +0000 UTC m=+156.201162065" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.685815 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" event={"ID":"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88","Type":"ContainerStarted","Data":"29e626b8c17193b8af3af3bae25eabcf87a14bb8cd3890303cb7d224f46f468f"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.685881 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" event={"ID":"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88","Type":"ContainerStarted","Data":"8e99ad2d868da6b69accd7124c41af978bbc3ecb53dc0829f7b1e7b21f75ac84"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.694584 5035 generic.go:334] "Generic (PLEG): container finished" podID="bb15b826-35be-4433-9e0f-88e013e8141d" containerID="8b576168f06689e429a7f7a277964f8090809394792acf2710f3551e0d721d58" exitCode=0 Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.695784 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" event={"ID":"bb15b826-35be-4433-9e0f-88e013e8141d","Type":"ContainerDied","Data":"8b576168f06689e429a7f7a277964f8090809394792acf2710f3551e0d721d58"} Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.708718 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.709132 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.209116842 +0000 UTC m=+156.731623099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.715613 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.718118 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9jm8b" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.728795 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7mkm9" podStartSLOduration=130.728775734 podStartE2EDuration="2m10.728775734s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.682438691 +0000 UTC m=+156.204944968" watchObservedRunningTime="2025-11-24 19:16:57.728775734 +0000 UTC m=+156.251281991" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.768469 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7p62f" podStartSLOduration=130.768453676 podStartE2EDuration="2m10.768453676s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.730689456 +0000 UTC m=+156.253195713" watchObservedRunningTime="2025-11-24 19:16:57.768453676 +0000 UTC m=+156.290959933" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.801823 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-zfmld" podStartSLOduration=130.801810718 podStartE2EDuration="2m10.801810718s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.800359679 +0000 UTC m=+156.322865946" watchObservedRunningTime="2025-11-24 19:16:57.801810718 +0000 UTC m=+156.324316975" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.809960 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.810751 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.31072275 +0000 UTC m=+156.833229057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.862912 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" podStartSLOduration=130.862897661 podStartE2EDuration="2m10.862897661s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.862245733 +0000 UTC m=+156.384751990" watchObservedRunningTime="2025-11-24 19:16:57.862897661 +0000 UTC m=+156.385403918" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.865176 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" podStartSLOduration=130.865168753 podStartE2EDuration="2m10.865168753s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.843979979 +0000 UTC m=+156.366486236" watchObservedRunningTime="2025-11-24 19:16:57.865168753 +0000 UTC m=+156.387675010" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.912947 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:57 crc kubenswrapper[5035]: E1124 19:16:57.913464 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.413446898 +0000 UTC m=+156.935953155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.949817 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" podStartSLOduration=117.949799311 podStartE2EDuration="1m57.949799311s" podCreationTimestamp="2025-11-24 19:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.948529956 +0000 UTC m=+156.471036213" watchObservedRunningTime="2025-11-24 19:16:57.949799311 +0000 UTC m=+156.472305568" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.950736 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" podStartSLOduration=130.950729736 podStartE2EDuration="2m10.950729736s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.923941872 +0000 UTC m=+156.446448129" watchObservedRunningTime="2025-11-24 19:16:57.950729736 +0000 UTC m=+156.473235993" Nov 24 19:16:57 crc kubenswrapper[5035]: I1124 19:16:57.987969 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-lwfhz" podStartSLOduration=130.987951412 podStartE2EDuration="2m10.987951412s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:57.985139646 +0000 UTC m=+156.507645913" watchObservedRunningTime="2025-11-24 19:16:57.987951412 +0000 UTC m=+156.510457669" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.014258 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.014764 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.514747277 +0000 UTC m=+157.037253534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.029805 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74lhf" podStartSLOduration=131.029792294 podStartE2EDuration="2m11.029792294s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.029048514 +0000 UTC m=+156.551554771" watchObservedRunningTime="2025-11-24 19:16:58.029792294 +0000 UTC m=+156.552298551" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.112877 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" podStartSLOduration=131.11286042 podStartE2EDuration="2m11.11286042s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.112558592 +0000 UTC m=+156.635064849" watchObservedRunningTime="2025-11-24 19:16:58.11286042 +0000 UTC m=+156.635366677" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.114529 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" podStartSLOduration=131.114522136 podStartE2EDuration="2m11.114522136s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.076374563 +0000 UTC m=+156.598880820" watchObservedRunningTime="2025-11-24 19:16:58.114522136 +0000 UTC m=+156.637028393" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.116184 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.116510 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.616499809 +0000 UTC m=+157.139006066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.219867 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.220472 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.7204552 +0000 UTC m=+157.242961457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.302329 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-wj8q6" podStartSLOduration=7.3023109139999995 podStartE2EDuration="7.302310914s" podCreationTimestamp="2025-11-24 19:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.267890603 +0000 UTC m=+156.790396860" watchObservedRunningTime="2025-11-24 19:16:58.302310914 +0000 UTC m=+156.824817171" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.303483 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sv2zq" podStartSLOduration=131.303475675 podStartE2EDuration="2m11.303475675s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.300993848 +0000 UTC m=+156.823500105" watchObservedRunningTime="2025-11-24 19:16:58.303475675 +0000 UTC m=+156.825981932" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.321408 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.327544 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.827527326 +0000 UTC m=+157.350033573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.424807 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.425367 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:58.925350601 +0000 UTC m=+157.447856858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.490743 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:16:58 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:16:58 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:16:58 crc kubenswrapper[5035]: healthz check failed Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.490799 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.526149 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.526507 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.026495236 +0000 UTC m=+157.549001493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.627335 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.627504 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.127477357 +0000 UTC m=+157.649983614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.627599 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.627877 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.127863197 +0000 UTC m=+157.650369454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.701410 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-24hv8" event={"ID":"f37e45f1-e73d-4bc0-904c-8ef5354ead61","Type":"ContainerStarted","Data":"b173af0d098ff632d334006966d878d4f190d49b7b779eba56d6a073216c7291"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.703587 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" event={"ID":"87a83e4e-6780-4f2a-9d77-bc87e906fd5f","Type":"ContainerStarted","Data":"070a86023bf261153be5be24947cd0d855c2d3c21cb71baa2481949de1630a5c"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.704160 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.705584 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" event={"ID":"facd9ded-06ed-439c-85ee-b5edd975c1f2","Type":"ContainerStarted","Data":"8082f648b0b33117a280366b71c3d581b7b24acb9251a0feaad84ca5a6fee048"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.707059 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" event={"ID":"6c737878-3aa7-4291-94dc-55a537dd0fae","Type":"ContainerStarted","Data":"d09564873609845be63193d0b4b09ec13cc44acf25aa99ec640a8d5a1a7d56f6"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.708474 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" event={"ID":"b08695f9-1ea8-4852-8b0a-0d548f000194","Type":"ContainerStarted","Data":"4dc905adf32e5a73fdce409c1474a2d9529cf5ea0130486ba81c494380eea2b4"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.710740 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" event={"ID":"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e","Type":"ContainerStarted","Data":"aba36979ec6f81e4ca1216a2fe2b4705e52eb8803bc497d9c6ade6130fe7ba47"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.710763 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" event={"ID":"0f7401bc-60a7-4cae-8e00-2f4ddb7e5e8e","Type":"ContainerStarted","Data":"435a8dd54e364ad39a36d09fc73b8edeb9e6d2c8277042f13804d9d73efb019b"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.713114 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" event={"ID":"3e1e4426-62ad-4454-bbc3-6e117657e33e","Type":"ContainerStarted","Data":"a136e4b01eab9f74201d840bcbdc2739bc07952efd8a97c9001e6c0c4560e9bd"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.714749 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" event={"ID":"dedcaa39-2f77-4754-ae51-69b47eb8eeb1","Type":"ContainerStarted","Data":"209212faaeee49941f1dd9f3c8d81d7bc6e21679907db17ccfb8826c253131e7"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.718682 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" event={"ID":"bb15b826-35be-4433-9e0f-88e013e8141d","Type":"ContainerStarted","Data":"db7f99382df0b20c22b32865c7295c161f77eeff47e46c44fe486dce1dfac0ec"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.718740 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.722020 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-csmtf" event={"ID":"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0","Type":"ContainerStarted","Data":"f158f4f86ea984f045d7fce07cd1d78d92bf0e50b23c9b54a2070170ea4d73da"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.722070 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-csmtf" event={"ID":"82de57d0-1931-4f1a-9ae0-d16aa81c7fd0","Type":"ContainerStarted","Data":"95879d382f34fba1835fd82778d2ecccdc0bbe8d88381aba8dbe3d52ff6e75fd"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.722119 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-csmtf" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.724977 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-47f49" event={"ID":"15eb6fde-d4d1-4ea4-b70f-f956105126a7","Type":"ContainerStarted","Data":"48467dbff3dd27ea0d206bb00d6cc437eb9a52df8917128e07491a1eed1a8ce6"} Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.726649 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" podStartSLOduration=131.726633698 podStartE2EDuration="2m11.726633698s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.72632033 +0000 UTC m=+157.248826587" watchObservedRunningTime="2025-11-24 19:16:58.726633698 +0000 UTC m=+157.249139955" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.727128 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-v8j4d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.727156 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-v8j4d" podUID="1d176339-2edc-4cf3-ad85-f4a0519f41c2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.727926 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.728150 5035 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q2jhw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.728203 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.728259 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.228244272 +0000 UTC m=+157.750750529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.733476 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.737263 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bvf2t" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.737752 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8b8fg" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.760254 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.818248 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" podStartSLOduration=132.818235175 podStartE2EDuration="2m12.818235175s" podCreationTimestamp="2025-11-24 19:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.774123383 +0000 UTC m=+157.296629640" watchObservedRunningTime="2025-11-24 19:16:58.818235175 +0000 UTC m=+157.340741432" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.818708 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g6ldj" podStartSLOduration=131.818703138 podStartE2EDuration="2m11.818703138s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.816332784 +0000 UTC m=+157.338839051" watchObservedRunningTime="2025-11-24 19:16:58.818703138 +0000 UTC m=+157.341209395" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.829184 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.838940 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.338925055 +0000 UTC m=+157.861431312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.857219 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-csmtf" podStartSLOduration=7.857202099 podStartE2EDuration="7.857202099s" podCreationTimestamp="2025-11-24 19:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.854830575 +0000 UTC m=+157.377336832" watchObservedRunningTime="2025-11-24 19:16:58.857202099 +0000 UTC m=+157.379708356" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.883889 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" podStartSLOduration=131.88387213 podStartE2EDuration="2m11.88387213s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.881801595 +0000 UTC m=+157.404307872" watchObservedRunningTime="2025-11-24 19:16:58.88387213 +0000 UTC m=+157.406378387" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.927580 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-2j9k6" podStartSLOduration=131.927565422 podStartE2EDuration="2m11.927565422s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.925137737 +0000 UTC m=+157.447643994" watchObservedRunningTime="2025-11-24 19:16:58.927565422 +0000 UTC m=+157.450071679" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.933557 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:58 crc kubenswrapper[5035]: E1124 19:16:58.934060 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.434042657 +0000 UTC m=+157.956548914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.942880 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-k7v8b" Nov 24 19:16:58 crc kubenswrapper[5035]: I1124 19:16:58.984495 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-v5t7x" podStartSLOduration=131.984474681 podStartE2EDuration="2m11.984474681s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:16:58.950940344 +0000 UTC m=+157.473446601" watchObservedRunningTime="2025-11-24 19:16:58.984474681 +0000 UTC m=+157.506980938" Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.035146 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.035567 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.535556082 +0000 UTC m=+158.058062339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.139950 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.140403 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.640379977 +0000 UTC m=+158.162886254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.241968 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.242400 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.742385615 +0000 UTC m=+158.264891872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.343491 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.343700 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.843674215 +0000 UTC m=+158.366180472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.343850 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.344179 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.844167538 +0000 UTC m=+158.366673795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.444831 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.444956 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.944929933 +0000 UTC m=+158.467436190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.445436 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.445736 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:16:59.945728954 +0000 UTC m=+158.468235211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.496680 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:16:59 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:16:59 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:16:59 crc kubenswrapper[5035]: healthz check failed Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.496733 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.546395 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.546811 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.046794908 +0000 UTC m=+158.569301165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.647526 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.647825 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.14781483 +0000 UTC m=+158.670321087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.748144 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.748313 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.248271685 +0000 UTC m=+158.770777942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.748468 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.748770 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.248762429 +0000 UTC m=+158.771268686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.752140 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" event={"ID":"6c737878-3aa7-4291-94dc-55a537dd0fae","Type":"ContainerStarted","Data":"d28d7aff9ef14843af775c1215c72e5ba5444b98d39705a7365da9385c9d82b9"} Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.752209 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" event={"ID":"6c737878-3aa7-4291-94dc-55a537dd0fae","Type":"ContainerStarted","Data":"b821118b4a3489acbd8aeb8219967f3e1945e0e98c3ee995e4cab6a4132313e4"} Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.753711 5035 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-q2jhw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.753757 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.849097 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.849251 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.349223446 +0000 UTC m=+158.871729703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.849617 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.851148 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.351134338 +0000 UTC m=+158.873640595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.868210 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8sczb"] Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.869119 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.873687 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.886423 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8sczb"] Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.950871 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:16:59 crc kubenswrapper[5035]: E1124 19:16:59.951080 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.451052 +0000 UTC m=+158.973558267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.951131 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-utilities\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.951254 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-catalog-content\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:16:59 crc kubenswrapper[5035]: I1124 19:16:59.951361 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf8zh\" (UniqueName: \"kubernetes.io/projected/0d029a23-831f-48f1-939e-de845c7a7c5b-kube-api-access-sf8zh\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.053048 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-catalog-content\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.053088 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf8zh\" (UniqueName: \"kubernetes.io/projected/0d029a23-831f-48f1-939e-de845c7a7c5b-kube-api-access-sf8zh\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.053157 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.053190 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-utilities\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: E1124 19:17:00.053569 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.553554532 +0000 UTC m=+159.076060789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.053568 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-catalog-content\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.053642 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-utilities\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.080005 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dlgkh"] Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.080885 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.097759 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.123848 5035 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.131637 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf8zh\" (UniqueName: \"kubernetes.io/projected/0d029a23-831f-48f1-939e-de845c7a7c5b-kube-api-access-sf8zh\") pod \"certified-operators-8sczb\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.145693 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlgkh"] Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.156154 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:17:00 crc kubenswrapper[5035]: E1124 19:17:00.156445 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.656428673 +0000 UTC m=+159.178934930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.180625 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.257956 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jklxv\" (UniqueName: \"kubernetes.io/projected/bc56ef71-4ba1-4378-b7ea-a97141be037e-kube-api-access-jklxv\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.258221 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.258242 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-utilities\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.258276 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-catalog-content\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: E1124 19:17:00.258659 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.758643147 +0000 UTC m=+159.281149404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.265244 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jczdc"] Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.266149 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.280484 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jczdc"] Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.359596 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.359834 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-catalog-content\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.359887 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jklxv\" (UniqueName: \"kubernetes.io/projected/bc56ef71-4ba1-4378-b7ea-a97141be037e-kube-api-access-jklxv\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.359929 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-utilities\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.359950 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bh8r\" (UniqueName: \"kubernetes.io/projected/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-kube-api-access-9bh8r\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.359983 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-utilities\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.360004 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-catalog-content\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: E1124 19:17:00.360416 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.860394269 +0000 UTC m=+159.382900526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.360453 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-catalog-content\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.360739 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-utilities\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.389837 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jklxv\" (UniqueName: \"kubernetes.io/projected/bc56ef71-4ba1-4378-b7ea-a97141be037e-kube-api-access-jklxv\") pod \"community-operators-dlgkh\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.392597 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.453637 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wjfb7"] Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.454585 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.460704 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.460742 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bh8r\" (UniqueName: \"kubernetes.io/projected/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-kube-api-access-9bh8r\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.460776 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-utilities\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.460805 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-catalog-content\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.461196 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-catalog-content\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.461421 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-utilities\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: E1124 19:17:00.461607 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:17:00.961590475 +0000 UTC m=+159.484096812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.482241 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wjfb7"] Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.484475 5035 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T19:17:00.123870913Z","Handler":null,"Name":""} Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.492209 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:17:00 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:17:00 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:17:00 crc kubenswrapper[5035]: healthz check failed Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.492256 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.500149 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bh8r\" (UniqueName: \"kubernetes.io/projected/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-kube-api-access-9bh8r\") pod \"certified-operators-jczdc\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.561547 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:17:00 crc kubenswrapper[5035]: E1124 19:17:00.561704 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 19:17:01.061676871 +0000 UTC m=+159.584183128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.561852 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-utilities\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.561889 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.561939 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7qxw\" (UniqueName: \"kubernetes.io/projected/833de199-4de2-4674-8709-e15c8cd9b40c-kube-api-access-n7qxw\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.561964 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-catalog-content\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: E1124 19:17:00.562178 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 19:17:01.062170335 +0000 UTC m=+159.584676592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zv4sz" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.575722 5035 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.576013 5035 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.607012 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.663841 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.664055 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7qxw\" (UniqueName: \"kubernetes.io/projected/833de199-4de2-4674-8709-e15c8cd9b40c-kube-api-access-n7qxw\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.664106 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-catalog-content\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.664162 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-utilities\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.664688 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-utilities\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.664887 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-catalog-content\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.695411 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7qxw\" (UniqueName: \"kubernetes.io/projected/833de199-4de2-4674-8709-e15c8cd9b40c-kube-api-access-n7qxw\") pod \"community-operators-wjfb7\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.696730 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.766053 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" event={"ID":"6c737878-3aa7-4291-94dc-55a537dd0fae","Type":"ContainerStarted","Data":"8ab7067aa3a4ad36b5cdbb17e7add704193c60769616faf1b85521bf179b5d1b"} Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.766985 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.767358 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.789855 5035 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.789888 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.806408 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vzvml" podStartSLOduration=9.806392549 podStartE2EDuration="9.806392549s" podCreationTimestamp="2025-11-24 19:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:17:00.805933257 +0000 UTC m=+159.328439514" watchObservedRunningTime="2025-11-24 19:17:00.806392549 +0000 UTC m=+159.328898806" Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.879080 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8sczb"] Nov 24 19:17:00 crc kubenswrapper[5035]: I1124 19:17:00.879732 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zv4sz\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.013398 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jczdc"] Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.023294 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlgkh"] Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.058530 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.184711 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wjfb7"] Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.395376 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zv4sz"] Nov 24 19:17:01 crc kubenswrapper[5035]: W1124 19:17:01.417163 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod807e010a_4744_4b1d_aa7c_9cd3360e19af.slice/crio-891281a9f0fe5e3b6547ce4fcae0da7ccdd0f6291c97c4189db59f5e41dcefa8 WatchSource:0}: Error finding container 891281a9f0fe5e3b6547ce4fcae0da7ccdd0f6291c97c4189db59f5e41dcefa8: Status 404 returned error can't find the container with id 891281a9f0fe5e3b6547ce4fcae0da7ccdd0f6291c97c4189db59f5e41dcefa8 Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.487477 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:17:01 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:17:01 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:17:01 crc kubenswrapper[5035]: healthz check failed Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.487515 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.578539 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.579150 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.581091 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.581528 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.631430 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.688801 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51cfcad3-1784-4f06-86bb-30612cf4308f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.688876 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51cfcad3-1784-4f06-86bb-30612cf4308f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.770738 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" event={"ID":"807e010a-4744-4b1d-aa7c-9cd3360e19af","Type":"ContainerStarted","Data":"fb6f9a780b5da6f25a60fc89d5e1d327ae8b2f53f39ddccb3f30e2bac5b468f6"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.771205 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.771313 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" event={"ID":"807e010a-4744-4b1d-aa7c-9cd3360e19af","Type":"ContainerStarted","Data":"891281a9f0fe5e3b6547ce4fcae0da7ccdd0f6291c97c4189db59f5e41dcefa8"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.772409 5035 generic.go:334] "Generic (PLEG): container finished" podID="833de199-4de2-4674-8709-e15c8cd9b40c" containerID="171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31" exitCode=0 Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.772461 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wjfb7" event={"ID":"833de199-4de2-4674-8709-e15c8cd9b40c","Type":"ContainerDied","Data":"171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.772477 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wjfb7" event={"ID":"833de199-4de2-4674-8709-e15c8cd9b40c","Type":"ContainerStarted","Data":"6535e26dd290799ebddec70a7d161bca55d26feaa83239c92654fda0eda35aee"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.773832 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.773856 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczdc" event={"ID":"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb","Type":"ContainerDied","Data":"4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.773831 5035 generic.go:334] "Generic (PLEG): container finished" podID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerID="4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd" exitCode=0 Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.775267 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczdc" event={"ID":"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb","Type":"ContainerStarted","Data":"cf94336a4f73718880bf4c049b81b99406652542edc46365eb5e58d59691ec2b"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.776729 5035 generic.go:334] "Generic (PLEG): container finished" podID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerID="39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188" exitCode=0 Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.776779 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8sczb" event={"ID":"0d029a23-831f-48f1-939e-de845c7a7c5b","Type":"ContainerDied","Data":"39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.776804 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8sczb" event={"ID":"0d029a23-831f-48f1-939e-de845c7a7c5b","Type":"ContainerStarted","Data":"528b81faf6b89f71620107de9c4d7ff943d9ae8bd7e047c0dfc56eb9c12d9c81"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.778606 5035 generic.go:334] "Generic (PLEG): container finished" podID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerID="36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f" exitCode=0 Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.779833 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlgkh" event={"ID":"bc56ef71-4ba1-4378-b7ea-a97141be037e","Type":"ContainerDied","Data":"36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.779854 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlgkh" event={"ID":"bc56ef71-4ba1-4378-b7ea-a97141be037e","Type":"ContainerStarted","Data":"e3322c80489bdb4e309d06b6713dc1c7ce9f39ef23811f1527c75ae3fe4b29af"} Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.789717 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51cfcad3-1784-4f06-86bb-30612cf4308f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.789803 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51cfcad3-1784-4f06-86bb-30612cf4308f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.790037 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51cfcad3-1784-4f06-86bb-30612cf4308f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.796962 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" podStartSLOduration=134.796944277 podStartE2EDuration="2m14.796944277s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:17:01.796652638 +0000 UTC m=+160.319158895" watchObservedRunningTime="2025-11-24 19:17:01.796944277 +0000 UTC m=+160.319450534" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.825913 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51cfcad3-1784-4f06-86bb-30612cf4308f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.850726 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-swgf2"] Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.851890 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.854874 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.872420 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-swgf2"] Nov 24 19:17:01 crc kubenswrapper[5035]: I1124 19:17:01.892597 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:01.992494 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-utilities\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:01.992547 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzxrn\" (UniqueName: \"kubernetes.io/projected/7c332553-06e8-4b00-9c70-0feffd93f281-kube-api-access-wzxrn\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:01.992784 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-catalog-content\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.094426 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-catalog-content\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.094481 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-utilities\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.094500 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzxrn\" (UniqueName: \"kubernetes.io/projected/7c332553-06e8-4b00-9c70-0feffd93f281-kube-api-access-wzxrn\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.095166 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-catalog-content\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.095389 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-utilities\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.123168 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzxrn\" (UniqueName: \"kubernetes.io/projected/7c332553-06e8-4b00-9c70-0feffd93f281-kube-api-access-wzxrn\") pod \"redhat-marketplace-swgf2\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.164537 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.227537 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.250585 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bl4fw"] Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.251717 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.261536 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl4fw"] Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.326810 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 19:17:02 crc kubenswrapper[5035]: W1124 19:17:02.339951 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod51cfcad3_1784_4f06_86bb_30612cf4308f.slice/crio-a6d97e0826584b03fc34a9be5d6382102f417964a24cde09d065eab0422f0dde WatchSource:0}: Error finding container a6d97e0826584b03fc34a9be5d6382102f417964a24cde09d065eab0422f0dde: Status 404 returned error can't find the container with id a6d97e0826584b03fc34a9be5d6382102f417964a24cde09d065eab0422f0dde Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.407681 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-utilities\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.407728 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-catalog-content\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.407795 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzhhb\" (UniqueName: \"kubernetes.io/projected/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-kube-api-access-pzhhb\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.413226 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-swgf2"] Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.489191 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:17:02 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:17:02 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:17:02 crc kubenswrapper[5035]: healthz check failed Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.489258 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.508465 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-utilities\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.508525 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-catalog-content\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.508574 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzhhb\" (UniqueName: \"kubernetes.io/projected/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-kube-api-access-pzhhb\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.508912 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-utilities\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.508928 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-catalog-content\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.532530 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzhhb\" (UniqueName: \"kubernetes.io/projected/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-kube-api-access-pzhhb\") pod \"redhat-marketplace-bl4fw\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.568461 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.655874 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6b8k5" Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.787982 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51cfcad3-1784-4f06-86bb-30612cf4308f","Type":"ContainerStarted","Data":"018b58ee22c70b13bbda51df7583e895f9d9794e0cf1d92a074710da8337de68"} Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.788356 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51cfcad3-1784-4f06-86bb-30612cf4308f","Type":"ContainerStarted","Data":"a6d97e0826584b03fc34a9be5d6382102f417964a24cde09d065eab0422f0dde"} Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.788372 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl4fw"] Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.792077 5035 generic.go:334] "Generic (PLEG): container finished" podID="7c332553-06e8-4b00-9c70-0feffd93f281" containerID="e37d4605c669733317a1c3ff682c102723e1a4c7a1188efae2904eff5e154307" exitCode=0 Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.792207 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swgf2" event={"ID":"7c332553-06e8-4b00-9c70-0feffd93f281","Type":"ContainerDied","Data":"e37d4605c669733317a1c3ff682c102723e1a4c7a1188efae2904eff5e154307"} Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.792235 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swgf2" event={"ID":"7c332553-06e8-4b00-9c70-0feffd93f281","Type":"ContainerStarted","Data":"f58c8ba11a756d6439a809819fd64dd3cce4ff1641e9ce1a314921a903b01128"} Nov 24 19:17:02 crc kubenswrapper[5035]: W1124 19:17:02.798579 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e41759d_ff0c_412d_a9a6_7593b3ac69f0.slice/crio-efd82be3e804a29422eabde7716b9167d4bfdef1e38d0d14499eb37e70684468 WatchSource:0}: Error finding container efd82be3e804a29422eabde7716b9167d4bfdef1e38d0d14499eb37e70684468: Status 404 returned error can't find the container with id efd82be3e804a29422eabde7716b9167d4bfdef1e38d0d14499eb37e70684468 Nov 24 19:17:02 crc kubenswrapper[5035]: I1124 19:17:02.807499 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.807482534 podStartE2EDuration="1.807482534s" podCreationTimestamp="2025-11-24 19:17:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:17:02.804117883 +0000 UTC m=+161.326624140" watchObservedRunningTime="2025-11-24 19:17:02.807482534 +0000 UTC m=+161.329988791" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.248174 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5ltvz"] Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.249570 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.256820 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5ltvz"] Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.259644 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.322981 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27kfq\" (UniqueName: \"kubernetes.io/projected/5efc44e8-4d11-4f01-b819-e56df5f4d88a-kube-api-access-27kfq\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.323071 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-catalog-content\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.323114 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-utilities\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.425651 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-catalog-content\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.425713 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-utilities\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.425743 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27kfq\" (UniqueName: \"kubernetes.io/projected/5efc44e8-4d11-4f01-b819-e56df5f4d88a-kube-api-access-27kfq\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.426304 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-catalog-content\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.426491 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-utilities\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.445192 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27kfq\" (UniqueName: \"kubernetes.io/projected/5efc44e8-4d11-4f01-b819-e56df5f4d88a-kube-api-access-27kfq\") pod \"redhat-operators-5ltvz\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.487728 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:17:03 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:17:03 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:17:03 crc kubenswrapper[5035]: healthz check failed Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.487795 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.515402 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.515488 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.523335 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.578171 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.658276 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nschj"] Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.659416 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.661805 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nschj"] Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.664594 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-v8j4d container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.664702 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-v8j4d" podUID="1d176339-2edc-4cf3-ad85-f4a0519f41c2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.665002 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-v8j4d container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.665096 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-v8j4d" podUID="1d176339-2edc-4cf3-ad85-f4a0519f41c2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.678426 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.678526 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.679479 5035 patch_prober.go:28] interesting pod/console-f9d7485db-58jzg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.679516 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-58jzg" podUID="2a0d45af-2aba-4830-a849-c2933f94fb83" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.729080 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94bc7\" (UniqueName: \"kubernetes.io/projected/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-kube-api-access-94bc7\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.729124 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-catalog-content\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.729162 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-utilities\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.807404 5035 generic.go:334] "Generic (PLEG): container finished" podID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerID="60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4" exitCode=0 Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.807510 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl4fw" event={"ID":"4e41759d-ff0c-412d-a9a6-7593b3ac69f0","Type":"ContainerDied","Data":"60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4"} Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.807538 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl4fw" event={"ID":"4e41759d-ff0c-412d-a9a6-7593b3ac69f0","Type":"ContainerStarted","Data":"efd82be3e804a29422eabde7716b9167d4bfdef1e38d0d14499eb37e70684468"} Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.809440 5035 generic.go:334] "Generic (PLEG): container finished" podID="0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" containerID="29e626b8c17193b8af3af3bae25eabcf87a14bb8cd3890303cb7d224f46f468f" exitCode=0 Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.809517 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" event={"ID":"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88","Type":"ContainerDied","Data":"29e626b8c17193b8af3af3bae25eabcf87a14bb8cd3890303cb7d224f46f468f"} Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.815461 5035 generic.go:334] "Generic (PLEG): container finished" podID="51cfcad3-1784-4f06-86bb-30612cf4308f" containerID="018b58ee22c70b13bbda51df7583e895f9d9794e0cf1d92a074710da8337de68" exitCode=0 Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.815964 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51cfcad3-1784-4f06-86bb-30612cf4308f","Type":"ContainerDied","Data":"018b58ee22c70b13bbda51df7583e895f9d9794e0cf1d92a074710da8337de68"} Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.824041 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5z8h5" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.830327 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-utilities\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.830493 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94bc7\" (UniqueName: \"kubernetes.io/projected/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-kube-api-access-94bc7\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.830526 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-catalog-content\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.832803 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-utilities\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.833062 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-catalog-content\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.850607 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5ltvz"] Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.855559 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94bc7\" (UniqueName: \"kubernetes.io/projected/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-kube-api-access-94bc7\") pod \"redhat-operators-nschj\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:03 crc kubenswrapper[5035]: I1124 19:17:03.996759 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.257269 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nschj"] Nov 24 19:17:04 crc kubenswrapper[5035]: W1124 19:17:04.264611 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6ccb2d_56f0_41a7_bc1b_5784948c3d28.slice/crio-9e99394aeebc74b9659b03147d5cb85dc58535495d09a08745e961c607ef5b60 WatchSource:0}: Error finding container 9e99394aeebc74b9659b03147d5cb85dc58535495d09a08745e961c607ef5b60: Status 404 returned error can't find the container with id 9e99394aeebc74b9659b03147d5cb85dc58535495d09a08745e961c607ef5b60 Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.419274 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.486420 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.491214 5035 patch_prober.go:28] interesting pod/router-default-5444994796-w8dwg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 19:17:04 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Nov 24 19:17:04 crc kubenswrapper[5035]: [+]process-running ok Nov 24 19:17:04 crc kubenswrapper[5035]: healthz check failed Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.491248 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w8dwg" podUID="03c8636a-40d2-4609-bfb9-808a76cb0f19" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.678829 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.680370 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.684021 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.684142 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.686658 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.746382 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9bed013-a0a6-4038-9160-b24f47142f4f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.746428 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f9bed013-a0a6-4038-9160-b24f47142f4f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.847919 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9bed013-a0a6-4038-9160-b24f47142f4f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.847972 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f9bed013-a0a6-4038-9160-b24f47142f4f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.848067 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f9bed013-a0a6-4038-9160-b24f47142f4f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.867097 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9bed013-a0a6-4038-9160-b24f47142f4f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.878525 5035 generic.go:334] "Generic (PLEG): container finished" podID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerID="24d2e74d00b063c17fae6933934961e3c96a01eb687fce58bc22ddf6b38e186a" exitCode=0 Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.878636 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ltvz" event={"ID":"5efc44e8-4d11-4f01-b819-e56df5f4d88a","Type":"ContainerDied","Data":"24d2e74d00b063c17fae6933934961e3c96a01eb687fce58bc22ddf6b38e186a"} Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.878678 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ltvz" event={"ID":"5efc44e8-4d11-4f01-b819-e56df5f4d88a","Type":"ContainerStarted","Data":"68435385d934664ad6fc977dbc28f1dc5230412b5fd5d67c348c109a7666b120"} Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.895446 5035 generic.go:334] "Generic (PLEG): container finished" podID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerID="e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548" exitCode=0 Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.895541 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nschj" event={"ID":"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28","Type":"ContainerDied","Data":"e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548"} Nov 24 19:17:04 crc kubenswrapper[5035]: I1124 19:17:04.895584 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nschj" event={"ID":"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28","Type":"ContainerStarted","Data":"9e99394aeebc74b9659b03147d5cb85dc58535495d09a08745e961c607ef5b60"} Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.003619 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.274157 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.355204 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51cfcad3-1784-4f06-86bb-30612cf4308f-kube-api-access\") pod \"51cfcad3-1784-4f06-86bb-30612cf4308f\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.356900 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51cfcad3-1784-4f06-86bb-30612cf4308f-kubelet-dir\") pod \"51cfcad3-1784-4f06-86bb-30612cf4308f\" (UID: \"51cfcad3-1784-4f06-86bb-30612cf4308f\") " Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.356958 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51cfcad3-1784-4f06-86bb-30612cf4308f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "51cfcad3-1784-4f06-86bb-30612cf4308f" (UID: "51cfcad3-1784-4f06-86bb-30612cf4308f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.357579 5035 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51cfcad3-1784-4f06-86bb-30612cf4308f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.378497 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51cfcad3-1784-4f06-86bb-30612cf4308f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "51cfcad3-1784-4f06-86bb-30612cf4308f" (UID: "51cfcad3-1784-4f06-86bb-30612cf4308f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.394748 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.441087 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.458235 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpbqk\" (UniqueName: \"kubernetes.io/projected/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-kube-api-access-cpbqk\") pod \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.458276 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-config-volume\") pod \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.458334 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-secret-volume\") pod \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\" (UID: \"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88\") " Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.458563 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51cfcad3-1784-4f06-86bb-30612cf4308f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.459170 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-config-volume" (OuterVolumeSpecName: "config-volume") pod "0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" (UID: "0d4ae5fb-edf2-4b2b-b462-60146b3e7a88"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.463067 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-kube-api-access-cpbqk" (OuterVolumeSpecName: "kube-api-access-cpbqk") pod "0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" (UID: "0d4ae5fb-edf2-4b2b-b462-60146b3e7a88"). InnerVolumeSpecName "kube-api-access-cpbqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.464949 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" (UID: "0d4ae5fb-edf2-4b2b-b462-60146b3e7a88"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.492751 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.497931 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-w8dwg" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.563403 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.563429 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpbqk\" (UniqueName: \"kubernetes.io/projected/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-kube-api-access-cpbqk\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.563438 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.907665 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.907749 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"51cfcad3-1784-4f06-86bb-30612cf4308f","Type":"ContainerDied","Data":"a6d97e0826584b03fc34a9be5d6382102f417964a24cde09d065eab0422f0dde"} Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.907790 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6d97e0826584b03fc34a9be5d6382102f417964a24cde09d065eab0422f0dde" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.920121 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.920959 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn" event={"ID":"0d4ae5fb-edf2-4b2b-b462-60146b3e7a88","Type":"ContainerDied","Data":"8e99ad2d868da6b69accd7124c41af978bbc3ecb53dc0829f7b1e7b21f75ac84"} Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.920997 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e99ad2d868da6b69accd7124c41af978bbc3ecb53dc0829f7b1e7b21f75ac84" Nov 24 19:17:05 crc kubenswrapper[5035]: I1124 19:17:05.922524 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f9bed013-a0a6-4038-9160-b24f47142f4f","Type":"ContainerStarted","Data":"2967c50b6eb1d6702fd6f916fbe203097c9502e7316d0e15b8714ce3de1d8c33"} Nov 24 19:17:06 crc kubenswrapper[5035]: I1124 19:17:06.936145 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f9bed013-a0a6-4038-9160-b24f47142f4f","Type":"ContainerStarted","Data":"309807aba412eb5e3db50b75df6b8fe1391d18125c3909c97ccdd197fcfe7dd3"} Nov 24 19:17:06 crc kubenswrapper[5035]: I1124 19:17:06.956112 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.956090572 podStartE2EDuration="2.956090572s" podCreationTimestamp="2025-11-24 19:17:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:17:06.95488465 +0000 UTC m=+165.477390907" watchObservedRunningTime="2025-11-24 19:17:06.956090572 +0000 UTC m=+165.478596829" Nov 24 19:17:07 crc kubenswrapper[5035]: I1124 19:17:07.948940 5035 generic.go:334] "Generic (PLEG): container finished" podID="f9bed013-a0a6-4038-9160-b24f47142f4f" containerID="309807aba412eb5e3db50b75df6b8fe1391d18125c3909c97ccdd197fcfe7dd3" exitCode=0 Nov 24 19:17:07 crc kubenswrapper[5035]: I1124 19:17:07.948993 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f9bed013-a0a6-4038-9160-b24f47142f4f","Type":"ContainerDied","Data":"309807aba412eb5e3db50b75df6b8fe1391d18125c3909c97ccdd197fcfe7dd3"} Nov 24 19:17:09 crc kubenswrapper[5035]: I1124 19:17:09.287884 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-csmtf" Nov 24 19:17:09 crc kubenswrapper[5035]: I1124 19:17:09.439552 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:17:09 crc kubenswrapper[5035]: I1124 19:17:09.445122 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b6ac9917-ee3c-4686-a9e9-a8ba98efea02-metrics-certs\") pod \"network-metrics-daemon-l2pnt\" (UID: \"b6ac9917-ee3c-4686-a9e9-a8ba98efea02\") " pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:17:09 crc kubenswrapper[5035]: I1124 19:17:09.454307 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2pnt" Nov 24 19:17:13 crc kubenswrapper[5035]: I1124 19:17:13.668963 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-v8j4d" Nov 24 19:17:13 crc kubenswrapper[5035]: I1124 19:17:13.696059 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:17:13 crc kubenswrapper[5035]: I1124 19:17:13.704707 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:17:14 crc kubenswrapper[5035]: I1124 19:17:14.445146 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:14 crc kubenswrapper[5035]: I1124 19:17:14.522675 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f9bed013-a0a6-4038-9160-b24f47142f4f-kubelet-dir\") pod \"f9bed013-a0a6-4038-9160-b24f47142f4f\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " Nov 24 19:17:14 crc kubenswrapper[5035]: I1124 19:17:14.522742 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9bed013-a0a6-4038-9160-b24f47142f4f-kube-api-access\") pod \"f9bed013-a0a6-4038-9160-b24f47142f4f\" (UID: \"f9bed013-a0a6-4038-9160-b24f47142f4f\") " Nov 24 19:17:14 crc kubenswrapper[5035]: I1124 19:17:14.524342 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9bed013-a0a6-4038-9160-b24f47142f4f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f9bed013-a0a6-4038-9160-b24f47142f4f" (UID: "f9bed013-a0a6-4038-9160-b24f47142f4f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:17:14 crc kubenswrapper[5035]: I1124 19:17:14.530884 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9bed013-a0a6-4038-9160-b24f47142f4f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f9bed013-a0a6-4038-9160-b24f47142f4f" (UID: "f9bed013-a0a6-4038-9160-b24f47142f4f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:17:14 crc kubenswrapper[5035]: I1124 19:17:14.624501 5035 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f9bed013-a0a6-4038-9160-b24f47142f4f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:14 crc kubenswrapper[5035]: I1124 19:17:14.624841 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f9bed013-a0a6-4038-9160-b24f47142f4f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:15 crc kubenswrapper[5035]: I1124 19:17:15.001830 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f9bed013-a0a6-4038-9160-b24f47142f4f","Type":"ContainerDied","Data":"2967c50b6eb1d6702fd6f916fbe203097c9502e7316d0e15b8714ce3de1d8c33"} Nov 24 19:17:15 crc kubenswrapper[5035]: I1124 19:17:15.002168 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2967c50b6eb1d6702fd6f916fbe203097c9502e7316d0e15b8714ce3de1d8c33" Nov 24 19:17:15 crc kubenswrapper[5035]: I1124 19:17:15.001975 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 19:17:15 crc kubenswrapper[5035]: I1124 19:17:15.234169 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:17:15 crc kubenswrapper[5035]: I1124 19:17:15.234374 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:17:16 crc kubenswrapper[5035]: I1124 19:17:16.277666 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-l2pnt"] Nov 24 19:17:17 crc kubenswrapper[5035]: I1124 19:17:17.018718 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" event={"ID":"b6ac9917-ee3c-4686-a9e9-a8ba98efea02","Type":"ContainerStarted","Data":"1d94203d942c5b8c0682765d308c4c76251812a4f84b68e2abfc28becf6e208a"} Nov 24 19:17:20 crc kubenswrapper[5035]: I1124 19:17:20.690333 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 19:17:21 crc kubenswrapper[5035]: I1124 19:17:21.068033 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:17:30 crc kubenswrapper[5035]: E1124 19:17:30.992781 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 24 19:17:30 crc kubenswrapper[5035]: E1124 19:17:30.993628 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wzxrn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-swgf2_openshift-marketplace(7c332553-06e8-4b00-9c70-0feffd93f281): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 19:17:30 crc kubenswrapper[5035]: E1124 19:17:30.994869 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-swgf2" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" Nov 24 19:17:32 crc kubenswrapper[5035]: E1124 19:17:32.411812 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 19:17:32 crc kubenswrapper[5035]: E1124 19:17:32.412300 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-27kfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5ltvz_openshift-marketplace(5efc44e8-4d11-4f01-b819-e56df5f4d88a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 19:17:32 crc kubenswrapper[5035]: E1124 19:17:32.413461 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5ltvz" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" Nov 24 19:17:33 crc kubenswrapper[5035]: I1124 19:17:33.113705 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" event={"ID":"b6ac9917-ee3c-4686-a9e9-a8ba98efea02","Type":"ContainerStarted","Data":"90330532ee810723136b37aa5231f2714329f7bf479896250256cce1f63d6f46"} Nov 24 19:17:33 crc kubenswrapper[5035]: E1124 19:17:33.689390 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-swgf2" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" Nov 24 19:17:33 crc kubenswrapper[5035]: E1124 19:17:33.690272 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5ltvz" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" Nov 24 19:17:33 crc kubenswrapper[5035]: E1124 19:17:33.747499 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 19:17:33 crc kubenswrapper[5035]: E1124 19:17:33.747642 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jklxv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-dlgkh_openshift-marketplace(bc56ef71-4ba1-4378-b7ea-a97141be037e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 19:17:33 crc kubenswrapper[5035]: E1124 19:17:33.748993 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-dlgkh" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" Nov 24 19:17:34 crc kubenswrapper[5035]: I1124 19:17:34.479619 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jhwk9" Nov 24 19:17:34 crc kubenswrapper[5035]: E1124 19:17:34.983563 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-dlgkh" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.057752 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.057896 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sf8zh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8sczb_openshift-marketplace(0d029a23-831f-48f1-939e-de845c7a7c5b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.059118 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8sczb" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.068148 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.068331 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9bh8r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jczdc_openshift-marketplace(2881c99f-084a-4c91-b4a6-8ebfddaaa4bb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.069467 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jczdc" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.087320 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.087502 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n7qxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wjfb7_openshift-marketplace(833de199-4de2-4674-8709-e15c8cd9b40c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.089326 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wjfb7" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.105212 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.105508 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-94bc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-nschj_openshift-marketplace(7a6ccb2d-56f0-41a7-bc1b-5784948c3d28): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.107952 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-nschj" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.127404 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-nschj" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.127538 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8sczb" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.127649 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jczdc" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" Nov 24 19:17:35 crc kubenswrapper[5035]: E1124 19:17:35.127892 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wjfb7" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" Nov 24 19:17:36 crc kubenswrapper[5035]: I1124 19:17:36.133480 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l2pnt" event={"ID":"b6ac9917-ee3c-4686-a9e9-a8ba98efea02","Type":"ContainerStarted","Data":"817274dc89836eebeb5f4e8208114de0bb9f67f420c457d8f0588f2a5ad8f2df"} Nov 24 19:17:36 crc kubenswrapper[5035]: I1124 19:17:36.136110 5035 generic.go:334] "Generic (PLEG): container finished" podID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerID="ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632" exitCode=0 Nov 24 19:17:36 crc kubenswrapper[5035]: I1124 19:17:36.136190 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl4fw" event={"ID":"4e41759d-ff0c-412d-a9a6-7593b3ac69f0","Type":"ContainerDied","Data":"ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632"} Nov 24 19:17:36 crc kubenswrapper[5035]: I1124 19:17:36.256023 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-l2pnt" podStartSLOduration=169.254511793 podStartE2EDuration="2m49.254511793s" podCreationTimestamp="2025-11-24 19:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:17:36.25182123 +0000 UTC m=+194.774327487" watchObservedRunningTime="2025-11-24 19:17:36.254511793 +0000 UTC m=+194.777018070" Nov 24 19:17:37 crc kubenswrapper[5035]: I1124 19:17:37.144054 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl4fw" event={"ID":"4e41759d-ff0c-412d-a9a6-7593b3ac69f0","Type":"ContainerStarted","Data":"7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0"} Nov 24 19:17:37 crc kubenswrapper[5035]: I1124 19:17:37.173489 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bl4fw" podStartSLOduration=2.429805307 podStartE2EDuration="35.173451933s" podCreationTimestamp="2025-11-24 19:17:02 +0000 UTC" firstStartedPulling="2025-11-24 19:17:03.811508326 +0000 UTC m=+162.334014583" lastFinishedPulling="2025-11-24 19:17:36.555154952 +0000 UTC m=+195.077661209" observedRunningTime="2025-11-24 19:17:37.166270449 +0000 UTC m=+195.688776706" watchObservedRunningTime="2025-11-24 19:17:37.173451933 +0000 UTC m=+195.695958240" Nov 24 19:17:42 crc kubenswrapper[5035]: I1124 19:17:42.569432 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:42 crc kubenswrapper[5035]: I1124 19:17:42.570025 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:42 crc kubenswrapper[5035]: I1124 19:17:42.719885 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:43 crc kubenswrapper[5035]: I1124 19:17:43.217579 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:43 crc kubenswrapper[5035]: I1124 19:17:43.254939 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl4fw"] Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.182740 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bl4fw" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="registry-server" containerID="cri-o://7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0" gracePeriod=2 Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.235401 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.235485 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.539368 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.643493 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-catalog-content\") pod \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.643574 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzhhb\" (UniqueName: \"kubernetes.io/projected/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-kube-api-access-pzhhb\") pod \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.643695 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-utilities\") pod \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\" (UID: \"4e41759d-ff0c-412d-a9a6-7593b3ac69f0\") " Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.644301 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-utilities" (OuterVolumeSpecName: "utilities") pod "4e41759d-ff0c-412d-a9a6-7593b3ac69f0" (UID: "4e41759d-ff0c-412d-a9a6-7593b3ac69f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.650136 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-kube-api-access-pzhhb" (OuterVolumeSpecName: "kube-api-access-pzhhb") pod "4e41759d-ff0c-412d-a9a6-7593b3ac69f0" (UID: "4e41759d-ff0c-412d-a9a6-7593b3ac69f0"). InnerVolumeSpecName "kube-api-access-pzhhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.667312 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e41759d-ff0c-412d-a9a6-7593b3ac69f0" (UID: "4e41759d-ff0c-412d-a9a6-7593b3ac69f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.745121 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.745158 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzhhb\" (UniqueName: \"kubernetes.io/projected/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-kube-api-access-pzhhb\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:45 crc kubenswrapper[5035]: I1124 19:17:45.745173 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e41759d-ff0c-412d-a9a6-7593b3ac69f0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.192851 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlgkh" event={"ID":"bc56ef71-4ba1-4378-b7ea-a97141be037e","Type":"ContainerStarted","Data":"aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8"} Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.196203 5035 generic.go:334] "Generic (PLEG): container finished" podID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerID="7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0" exitCode=0 Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.196260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl4fw" event={"ID":"4e41759d-ff0c-412d-a9a6-7593b3ac69f0","Type":"ContainerDied","Data":"7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0"} Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.196311 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bl4fw" event={"ID":"4e41759d-ff0c-412d-a9a6-7593b3ac69f0","Type":"ContainerDied","Data":"efd82be3e804a29422eabde7716b9167d4bfdef1e38d0d14499eb37e70684468"} Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.196333 5035 scope.go:117] "RemoveContainer" containerID="7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.196507 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bl4fw" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.210922 5035 scope.go:117] "RemoveContainer" containerID="ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.263423 5035 scope.go:117] "RemoveContainer" containerID="60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.272676 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl4fw"] Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.279364 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bl4fw"] Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.289007 5035 scope.go:117] "RemoveContainer" containerID="7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0" Nov 24 19:17:46 crc kubenswrapper[5035]: E1124 19:17:46.289557 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0\": container with ID starting with 7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0 not found: ID does not exist" containerID="7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.289641 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0"} err="failed to get container status \"7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0\": rpc error: code = NotFound desc = could not find container \"7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0\": container with ID starting with 7f9547942ad74e97867f24d9228c5820797d61903f68c887d8344e8db06c50a0 not found: ID does not exist" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.289691 5035 scope.go:117] "RemoveContainer" containerID="ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632" Nov 24 19:17:46 crc kubenswrapper[5035]: E1124 19:17:46.290002 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632\": container with ID starting with ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632 not found: ID does not exist" containerID="ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.290034 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632"} err="failed to get container status \"ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632\": rpc error: code = NotFound desc = could not find container \"ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632\": container with ID starting with ce44df8cb5c065da2f880a369401d0a180206a621f4afb66c5ba7bc7dac9e632 not found: ID does not exist" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.290060 5035 scope.go:117] "RemoveContainer" containerID="60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4" Nov 24 19:17:46 crc kubenswrapper[5035]: E1124 19:17:46.290392 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4\": container with ID starting with 60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4 not found: ID does not exist" containerID="60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4" Nov 24 19:17:46 crc kubenswrapper[5035]: I1124 19:17:46.290414 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4"} err="failed to get container status \"60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4\": rpc error: code = NotFound desc = could not find container \"60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4\": container with ID starting with 60349bc7d9b1a6b8d845bfcd00ec6f98f5ec0d6f61415c627f242cd9c4d6a6b4 not found: ID does not exist" Nov 24 19:17:47 crc kubenswrapper[5035]: I1124 19:17:47.203722 5035 generic.go:334] "Generic (PLEG): container finished" podID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerID="aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8" exitCode=0 Nov 24 19:17:47 crc kubenswrapper[5035]: I1124 19:17:47.203859 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlgkh" event={"ID":"bc56ef71-4ba1-4378-b7ea-a97141be037e","Type":"ContainerDied","Data":"aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8"} Nov 24 19:17:47 crc kubenswrapper[5035]: I1124 19:17:47.209321 5035 generic.go:334] "Generic (PLEG): container finished" podID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerID="ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1" exitCode=0 Nov 24 19:17:47 crc kubenswrapper[5035]: I1124 19:17:47.209354 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8sczb" event={"ID":"0d029a23-831f-48f1-939e-de845c7a7c5b","Type":"ContainerDied","Data":"ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1"} Nov 24 19:17:48 crc kubenswrapper[5035]: I1124 19:17:48.205800 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" path="/var/lib/kubelet/pods/4e41759d-ff0c-412d-a9a6-7593b3ac69f0/volumes" Nov 24 19:17:48 crc kubenswrapper[5035]: I1124 19:17:48.216404 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlgkh" event={"ID":"bc56ef71-4ba1-4378-b7ea-a97141be037e","Type":"ContainerStarted","Data":"d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3"} Nov 24 19:17:48 crc kubenswrapper[5035]: I1124 19:17:48.218092 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ltvz" event={"ID":"5efc44e8-4d11-4f01-b819-e56df5f4d88a","Type":"ContainerStarted","Data":"78af914582f6e3c310f4d8f84ceeb54ad798c4758f3ffc5aa7525c0a248ce8e1"} Nov 24 19:17:48 crc kubenswrapper[5035]: I1124 19:17:48.219604 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8sczb" event={"ID":"0d029a23-831f-48f1-939e-de845c7a7c5b","Type":"ContainerStarted","Data":"17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd"} Nov 24 19:17:48 crc kubenswrapper[5035]: I1124 19:17:48.294781 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8sczb" podStartSLOduration=3.400260825 podStartE2EDuration="49.294765217s" podCreationTimestamp="2025-11-24 19:16:59 +0000 UTC" firstStartedPulling="2025-11-24 19:17:01.777797449 +0000 UTC m=+160.300303706" lastFinishedPulling="2025-11-24 19:17:47.672301851 +0000 UTC m=+206.194808098" observedRunningTime="2025-11-24 19:17:48.293251136 +0000 UTC m=+206.815757393" watchObservedRunningTime="2025-11-24 19:17:48.294765217 +0000 UTC m=+206.817271474" Nov 24 19:17:49 crc kubenswrapper[5035]: I1124 19:17:49.230004 5035 generic.go:334] "Generic (PLEG): container finished" podID="7c332553-06e8-4b00-9c70-0feffd93f281" containerID="5df1613c5757f61dd55605151cc1a3872380dff9977d497cbb36c3389c91fc50" exitCode=0 Nov 24 19:17:49 crc kubenswrapper[5035]: I1124 19:17:49.230098 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swgf2" event={"ID":"7c332553-06e8-4b00-9c70-0feffd93f281","Type":"ContainerDied","Data":"5df1613c5757f61dd55605151cc1a3872380dff9977d497cbb36c3389c91fc50"} Nov 24 19:17:49 crc kubenswrapper[5035]: I1124 19:17:49.232843 5035 generic.go:334] "Generic (PLEG): container finished" podID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerID="78af914582f6e3c310f4d8f84ceeb54ad798c4758f3ffc5aa7525c0a248ce8e1" exitCode=0 Nov 24 19:17:49 crc kubenswrapper[5035]: I1124 19:17:49.232929 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ltvz" event={"ID":"5efc44e8-4d11-4f01-b819-e56df5f4d88a","Type":"ContainerDied","Data":"78af914582f6e3c310f4d8f84ceeb54ad798c4758f3ffc5aa7525c0a248ce8e1"} Nov 24 19:17:49 crc kubenswrapper[5035]: I1124 19:17:49.238017 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nschj" event={"ID":"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28","Type":"ContainerStarted","Data":"fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751"} Nov 24 19:17:49 crc kubenswrapper[5035]: I1124 19:17:49.260224 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dlgkh" podStartSLOduration=3.337127224 podStartE2EDuration="49.260203234s" podCreationTimestamp="2025-11-24 19:17:00 +0000 UTC" firstStartedPulling="2025-11-24 19:17:01.780783479 +0000 UTC m=+160.303289726" lastFinishedPulling="2025-11-24 19:17:47.703859479 +0000 UTC m=+206.226365736" observedRunningTime="2025-11-24 19:17:48.334188788 +0000 UTC m=+206.856695045" watchObservedRunningTime="2025-11-24 19:17:49.260203234 +0000 UTC m=+207.782709481" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.181597 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.181889 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.232284 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.251353 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swgf2" event={"ID":"7c332553-06e8-4b00-9c70-0feffd93f281","Type":"ContainerStarted","Data":"dc525aa8e4bf2a4f4889de7d87e48b7da78f13f3a232640cb7cfe3556a214c67"} Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.254710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ltvz" event={"ID":"5efc44e8-4d11-4f01-b819-e56df5f4d88a","Type":"ContainerStarted","Data":"fdd1ffb8521e97699a8b5d667d66fe5f7824eb8404650bc74c712d1e265f610d"} Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.256653 5035 generic.go:334] "Generic (PLEG): container finished" podID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerID="fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751" exitCode=0 Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.256732 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nschj" event={"ID":"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28","Type":"ContainerDied","Data":"fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751"} Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.317647 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-swgf2" podStartSLOduration=2.474298647 podStartE2EDuration="49.31762648s" podCreationTimestamp="2025-11-24 19:17:01 +0000 UTC" firstStartedPulling="2025-11-24 19:17:02.794886343 +0000 UTC m=+161.317392600" lastFinishedPulling="2025-11-24 19:17:49.638214176 +0000 UTC m=+208.160720433" observedRunningTime="2025-11-24 19:17:50.317154976 +0000 UTC m=+208.839661253" watchObservedRunningTime="2025-11-24 19:17:50.31762648 +0000 UTC m=+208.840132737" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.337516 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5ltvz" podStartSLOduration=2.560173294 podStartE2EDuration="47.33749461s" podCreationTimestamp="2025-11-24 19:17:03 +0000 UTC" firstStartedPulling="2025-11-24 19:17:04.887480852 +0000 UTC m=+163.409987109" lastFinishedPulling="2025-11-24 19:17:49.664802168 +0000 UTC m=+208.187308425" observedRunningTime="2025-11-24 19:17:50.336829821 +0000 UTC m=+208.859336098" watchObservedRunningTime="2025-11-24 19:17:50.33749461 +0000 UTC m=+208.860000887" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.393125 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.393420 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:50 crc kubenswrapper[5035]: I1124 19:17:50.430088 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:51 crc kubenswrapper[5035]: I1124 19:17:51.264259 5035 generic.go:334] "Generic (PLEG): container finished" podID="833de199-4de2-4674-8709-e15c8cd9b40c" containerID="4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc" exitCode=0 Nov 24 19:17:51 crc kubenswrapper[5035]: I1124 19:17:51.264324 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wjfb7" event={"ID":"833de199-4de2-4674-8709-e15c8cd9b40c","Type":"ContainerDied","Data":"4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc"} Nov 24 19:17:51 crc kubenswrapper[5035]: I1124 19:17:51.265777 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nschj" event={"ID":"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28","Type":"ContainerStarted","Data":"66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2"} Nov 24 19:17:51 crc kubenswrapper[5035]: I1124 19:17:51.298023 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nschj" podStartSLOduration=2.503342718 podStartE2EDuration="48.298004252s" podCreationTimestamp="2025-11-24 19:17:03 +0000 UTC" firstStartedPulling="2025-11-24 19:17:04.896924608 +0000 UTC m=+163.419430865" lastFinishedPulling="2025-11-24 19:17:50.691586142 +0000 UTC m=+209.214092399" observedRunningTime="2025-11-24 19:17:51.296760428 +0000 UTC m=+209.819266685" watchObservedRunningTime="2025-11-24 19:17:51.298004252 +0000 UTC m=+209.820510519" Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.165332 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.165653 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.211391 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.273763 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wjfb7" event={"ID":"833de199-4de2-4674-8709-e15c8cd9b40c","Type":"ContainerStarted","Data":"1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418"} Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.275463 5035 generic.go:334] "Generic (PLEG): container finished" podID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerID="7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930" exitCode=0 Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.275545 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczdc" event={"ID":"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb","Type":"ContainerDied","Data":"7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930"} Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.296387 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wjfb7" podStartSLOduration=2.399146585 podStartE2EDuration="52.296369873s" podCreationTimestamp="2025-11-24 19:17:00 +0000 UTC" firstStartedPulling="2025-11-24 19:17:01.773582395 +0000 UTC m=+160.296088652" lastFinishedPulling="2025-11-24 19:17:51.670805683 +0000 UTC m=+210.193311940" observedRunningTime="2025-11-24 19:17:52.294690728 +0000 UTC m=+210.817196995" watchObservedRunningTime="2025-11-24 19:17:52.296369873 +0000 UTC m=+210.818876130" Nov 24 19:17:52 crc kubenswrapper[5035]: I1124 19:17:52.311498 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:17:53 crc kubenswrapper[5035]: I1124 19:17:53.282987 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczdc" event={"ID":"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb","Type":"ContainerStarted","Data":"bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354"} Nov 24 19:17:53 crc kubenswrapper[5035]: I1124 19:17:53.305575 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jczdc" podStartSLOduration=2.190854714 podStartE2EDuration="53.305548528s" podCreationTimestamp="2025-11-24 19:17:00 +0000 UTC" firstStartedPulling="2025-11-24 19:17:01.774825959 +0000 UTC m=+160.297332216" lastFinishedPulling="2025-11-24 19:17:52.889519773 +0000 UTC m=+211.412026030" observedRunningTime="2025-11-24 19:17:53.304461239 +0000 UTC m=+211.826967486" watchObservedRunningTime="2025-11-24 19:17:53.305548528 +0000 UTC m=+211.828054785" Nov 24 19:17:53 crc kubenswrapper[5035]: I1124 19:17:53.579177 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:53 crc kubenswrapper[5035]: I1124 19:17:53.579220 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:17:53 crc kubenswrapper[5035]: I1124 19:17:53.997454 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:53 crc kubenswrapper[5035]: I1124 19:17:53.997846 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:17:54 crc kubenswrapper[5035]: I1124 19:17:54.621865 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5ltvz" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="registry-server" probeResult="failure" output=< Nov 24 19:17:54 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 19:17:54 crc kubenswrapper[5035]: > Nov 24 19:17:55 crc kubenswrapper[5035]: I1124 19:17:55.033333 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nschj" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="registry-server" probeResult="failure" output=< Nov 24 19:17:55 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 19:17:55 crc kubenswrapper[5035]: > Nov 24 19:18:00 crc kubenswrapper[5035]: I1124 19:18:00.246852 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:18:00 crc kubenswrapper[5035]: I1124 19:18:00.608103 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:18:00 crc kubenswrapper[5035]: I1124 19:18:00.608184 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:18:00 crc kubenswrapper[5035]: I1124 19:18:00.668155 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:18:00 crc kubenswrapper[5035]: I1124 19:18:00.783676 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:18:00 crc kubenswrapper[5035]: I1124 19:18:00.784103 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:18:00 crc kubenswrapper[5035]: I1124 19:18:00.841441 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:18:01 crc kubenswrapper[5035]: I1124 19:18:01.383583 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:18:01 crc kubenswrapper[5035]: I1124 19:18:01.391031 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:18:02 crc kubenswrapper[5035]: I1124 19:18:02.090497 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jczdc"] Nov 24 19:18:02 crc kubenswrapper[5035]: I1124 19:18:02.217839 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:18:03 crc kubenswrapper[5035]: I1124 19:18:03.090987 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wjfb7"] Nov 24 19:18:03 crc kubenswrapper[5035]: I1124 19:18:03.343557 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jczdc" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="registry-server" containerID="cri-o://bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354" gracePeriod=2 Nov 24 19:18:03 crc kubenswrapper[5035]: I1124 19:18:03.343660 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wjfb7" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="registry-server" containerID="cri-o://1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418" gracePeriod=2 Nov 24 19:18:03 crc kubenswrapper[5035]: I1124 19:18:03.620532 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:18:03 crc kubenswrapper[5035]: I1124 19:18:03.656969 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:18:04 crc kubenswrapper[5035]: I1124 19:18:04.037247 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:18:04 crc kubenswrapper[5035]: I1124 19:18:04.073495 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.135187 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.140579 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.218555 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-utilities\") pod \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.218649 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7qxw\" (UniqueName: \"kubernetes.io/projected/833de199-4de2-4674-8709-e15c8cd9b40c-kube-api-access-n7qxw\") pod \"833de199-4de2-4674-8709-e15c8cd9b40c\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.218715 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-utilities\") pod \"833de199-4de2-4674-8709-e15c8cd9b40c\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.218790 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bh8r\" (UniqueName: \"kubernetes.io/projected/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-kube-api-access-9bh8r\") pod \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.218817 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-catalog-content\") pod \"833de199-4de2-4674-8709-e15c8cd9b40c\" (UID: \"833de199-4de2-4674-8709-e15c8cd9b40c\") " Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.218856 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-catalog-content\") pod \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\" (UID: \"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb\") " Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.220061 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-utilities" (OuterVolumeSpecName: "utilities") pod "833de199-4de2-4674-8709-e15c8cd9b40c" (UID: "833de199-4de2-4674-8709-e15c8cd9b40c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.220275 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.219282 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-utilities" (OuterVolumeSpecName: "utilities") pod "2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" (UID: "2881c99f-084a-4c91-b4a6-8ebfddaaa4bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.232501 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-kube-api-access-9bh8r" (OuterVolumeSpecName: "kube-api-access-9bh8r") pod "2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" (UID: "2881c99f-084a-4c91-b4a6-8ebfddaaa4bb"). InnerVolumeSpecName "kube-api-access-9bh8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.232847 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833de199-4de2-4674-8709-e15c8cd9b40c-kube-api-access-n7qxw" (OuterVolumeSpecName: "kube-api-access-n7qxw") pod "833de199-4de2-4674-8709-e15c8cd9b40c" (UID: "833de199-4de2-4674-8709-e15c8cd9b40c"). InnerVolumeSpecName "kube-api-access-n7qxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.269713 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" (UID: "2881c99f-084a-4c91-b4a6-8ebfddaaa4bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.282116 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "833de199-4de2-4674-8709-e15c8cd9b40c" (UID: "833de199-4de2-4674-8709-e15c8cd9b40c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.321772 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.321805 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7qxw\" (UniqueName: \"kubernetes.io/projected/833de199-4de2-4674-8709-e15c8cd9b40c-kube-api-access-n7qxw\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.321816 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bh8r\" (UniqueName: \"kubernetes.io/projected/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-kube-api-access-9bh8r\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.321825 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/833de199-4de2-4674-8709-e15c8cd9b40c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.321834 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.360929 5035 generic.go:334] "Generic (PLEG): container finished" podID="833de199-4de2-4674-8709-e15c8cd9b40c" containerID="1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418" exitCode=0 Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.360995 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wjfb7" event={"ID":"833de199-4de2-4674-8709-e15c8cd9b40c","Type":"ContainerDied","Data":"1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418"} Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.361021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wjfb7" event={"ID":"833de199-4de2-4674-8709-e15c8cd9b40c","Type":"ContainerDied","Data":"6535e26dd290799ebddec70a7d161bca55d26feaa83239c92654fda0eda35aee"} Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.361044 5035 scope.go:117] "RemoveContainer" containerID="1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.361051 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wjfb7" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.363157 5035 generic.go:334] "Generic (PLEG): container finished" podID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerID="bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354" exitCode=0 Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.363184 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczdc" event={"ID":"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb","Type":"ContainerDied","Data":"bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354"} Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.363206 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczdc" event={"ID":"2881c99f-084a-4c91-b4a6-8ebfddaaa4bb","Type":"ContainerDied","Data":"cf94336a4f73718880bf4c049b81b99406652542edc46365eb5e58d59691ec2b"} Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.363237 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczdc" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.382932 5035 scope.go:117] "RemoveContainer" containerID="4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.396330 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jczdc"] Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.398822 5035 scope.go:117] "RemoveContainer" containerID="171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.399152 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jczdc"] Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.408005 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wjfb7"] Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.410495 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wjfb7"] Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.418079 5035 scope.go:117] "RemoveContainer" containerID="1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418" Nov 24 19:18:06 crc kubenswrapper[5035]: E1124 19:18:06.418594 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418\": container with ID starting with 1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418 not found: ID does not exist" containerID="1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.418626 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418"} err="failed to get container status \"1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418\": rpc error: code = NotFound desc = could not find container \"1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418\": container with ID starting with 1de11958d68292924c7b710390aa1334e1760cd3d2ea0b02867fd9b6563dc418 not found: ID does not exist" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.418652 5035 scope.go:117] "RemoveContainer" containerID="4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc" Nov 24 19:18:06 crc kubenswrapper[5035]: E1124 19:18:06.419183 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc\": container with ID starting with 4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc not found: ID does not exist" containerID="4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.419218 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc"} err="failed to get container status \"4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc\": rpc error: code = NotFound desc = could not find container \"4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc\": container with ID starting with 4cfb17f7bc5029816348bd2b9b17b288f4b2d457b6ffd3973c0b909176dfd6bc not found: ID does not exist" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.419245 5035 scope.go:117] "RemoveContainer" containerID="171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31" Nov 24 19:18:06 crc kubenswrapper[5035]: E1124 19:18:06.419604 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31\": container with ID starting with 171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31 not found: ID does not exist" containerID="171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.419798 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31"} err="failed to get container status \"171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31\": rpc error: code = NotFound desc = could not find container \"171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31\": container with ID starting with 171ab8bb64c1e1a3bef2ac46333be560f180e808086f7d7526ba12dfaf631e31 not found: ID does not exist" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.419818 5035 scope.go:117] "RemoveContainer" containerID="bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.436014 5035 scope.go:117] "RemoveContainer" containerID="7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.449217 5035 scope.go:117] "RemoveContainer" containerID="4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.468145 5035 scope.go:117] "RemoveContainer" containerID="bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354" Nov 24 19:18:06 crc kubenswrapper[5035]: E1124 19:18:06.468590 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354\": container with ID starting with bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354 not found: ID does not exist" containerID="bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.468622 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354"} err="failed to get container status \"bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354\": rpc error: code = NotFound desc = could not find container \"bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354\": container with ID starting with bd013d4432d4a6fa88b8d9df162f16903d0133afe2d1a469d5e64d70d8cb2354 not found: ID does not exist" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.468643 5035 scope.go:117] "RemoveContainer" containerID="7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930" Nov 24 19:18:06 crc kubenswrapper[5035]: E1124 19:18:06.469135 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930\": container with ID starting with 7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930 not found: ID does not exist" containerID="7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.469164 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930"} err="failed to get container status \"7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930\": rpc error: code = NotFound desc = could not find container \"7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930\": container with ID starting with 7cbdc76b1ea259ac01aa093587b4df604b65d263a862b49b605b391d6a099930 not found: ID does not exist" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.469183 5035 scope.go:117] "RemoveContainer" containerID="4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd" Nov 24 19:18:06 crc kubenswrapper[5035]: E1124 19:18:06.469658 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd\": container with ID starting with 4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd not found: ID does not exist" containerID="4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd" Nov 24 19:18:06 crc kubenswrapper[5035]: I1124 19:18:06.469683 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd"} err="failed to get container status \"4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd\": rpc error: code = NotFound desc = could not find container \"4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd\": container with ID starting with 4675afc8b984e938a69747af7cb52d5785cdb826b96b4c6e1a7570b21ea436cd not found: ID does not exist" Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.487776 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nschj"] Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.488095 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nschj" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="registry-server" containerID="cri-o://66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2" gracePeriod=2 Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.904472 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.940611 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-utilities\") pod \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.940709 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-catalog-content\") pod \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.940798 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94bc7\" (UniqueName: \"kubernetes.io/projected/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-kube-api-access-94bc7\") pod \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\" (UID: \"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28\") " Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.942039 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-utilities" (OuterVolumeSpecName: "utilities") pod "7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" (UID: "7a6ccb2d-56f0-41a7-bc1b-5784948c3d28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:18:07 crc kubenswrapper[5035]: I1124 19:18:07.946472 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-kube-api-access-94bc7" (OuterVolumeSpecName: "kube-api-access-94bc7") pod "7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" (UID: "7a6ccb2d-56f0-41a7-bc1b-5784948c3d28"). InnerVolumeSpecName "kube-api-access-94bc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.041460 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" (UID: "7a6ccb2d-56f0-41a7-bc1b-5784948c3d28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.041903 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94bc7\" (UniqueName: \"kubernetes.io/projected/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-kube-api-access-94bc7\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.041937 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.041949 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.205001 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" path="/var/lib/kubelet/pods/2881c99f-084a-4c91-b4a6-8ebfddaaa4bb/volumes" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.205762 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" path="/var/lib/kubelet/pods/833de199-4de2-4674-8709-e15c8cd9b40c/volumes" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.387079 5035 generic.go:334] "Generic (PLEG): container finished" podID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerID="66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2" exitCode=0 Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.387124 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nschj" event={"ID":"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28","Type":"ContainerDied","Data":"66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2"} Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.387155 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nschj" event={"ID":"7a6ccb2d-56f0-41a7-bc1b-5784948c3d28","Type":"ContainerDied","Data":"9e99394aeebc74b9659b03147d5cb85dc58535495d09a08745e961c607ef5b60"} Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.387176 5035 scope.go:117] "RemoveContainer" containerID="66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.387184 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nschj" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.408798 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nschj"] Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.412413 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nschj"] Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.417556 5035 scope.go:117] "RemoveContainer" containerID="fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.433703 5035 scope.go:117] "RemoveContainer" containerID="e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.472561 5035 scope.go:117] "RemoveContainer" containerID="66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2" Nov 24 19:18:08 crc kubenswrapper[5035]: E1124 19:18:08.473170 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2\": container with ID starting with 66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2 not found: ID does not exist" containerID="66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.473219 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2"} err="failed to get container status \"66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2\": rpc error: code = NotFound desc = could not find container \"66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2\": container with ID starting with 66daab08c76e4b465674079009f703e996ab5ca9f54f6c523c56aa585fe42bc2 not found: ID does not exist" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.473252 5035 scope.go:117] "RemoveContainer" containerID="fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751" Nov 24 19:18:08 crc kubenswrapper[5035]: E1124 19:18:08.473550 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751\": container with ID starting with fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751 not found: ID does not exist" containerID="fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.473573 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751"} err="failed to get container status \"fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751\": rpc error: code = NotFound desc = could not find container \"fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751\": container with ID starting with fe8ab441a239ec582ccc66c250e4ee7e5cd1835d5d7a476f9a35dd1a487fb751 not found: ID does not exist" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.473591 5035 scope.go:117] "RemoveContainer" containerID="e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548" Nov 24 19:18:08 crc kubenswrapper[5035]: E1124 19:18:08.474478 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548\": container with ID starting with e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548 not found: ID does not exist" containerID="e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548" Nov 24 19:18:08 crc kubenswrapper[5035]: I1124 19:18:08.474553 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548"} err="failed to get container status \"e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548\": rpc error: code = NotFound desc = could not find container \"e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548\": container with ID starting with e5bba0720d9fd1efbfea17ab089c3f0c9cd609cd8de050a7141fd146adab1548 not found: ID does not exist" Nov 24 19:18:10 crc kubenswrapper[5035]: I1124 19:18:10.207915 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" path="/var/lib/kubelet/pods/7a6ccb2d-56f0-41a7-bc1b-5784948c3d28/volumes" Nov 24 19:18:12 crc kubenswrapper[5035]: I1124 19:18:12.498871 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qfxzs"] Nov 24 19:18:15 crc kubenswrapper[5035]: I1124 19:18:15.234764 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:18:15 crc kubenswrapper[5035]: I1124 19:18:15.235154 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:18:15 crc kubenswrapper[5035]: I1124 19:18:15.235214 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:18:15 crc kubenswrapper[5035]: I1124 19:18:15.235898 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:18:15 crc kubenswrapper[5035]: I1124 19:18:15.235965 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3" gracePeriod=600 Nov 24 19:18:16 crc kubenswrapper[5035]: I1124 19:18:16.446644 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3" exitCode=0 Nov 24 19:18:16 crc kubenswrapper[5035]: I1124 19:18:16.446772 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3"} Nov 24 19:18:16 crc kubenswrapper[5035]: I1124 19:18:16.447035 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"71af2052370dcb5cad38af3277bc18c2dd745f2273be300ec615c3e8cdb031fa"} Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.522733 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" podUID="fb6648af-1ab7-419e-bdb3-8f696aaabd80" containerName="oauth-openshift" containerID="cri-o://6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de" gracePeriod=15 Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.887203 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.934379 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6644f974c8-r8xlb"] Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.934865 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51cfcad3-1784-4f06-86bb-30612cf4308f" containerName="pruner" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.934936 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="51cfcad3-1784-4f06-86bb-30612cf4308f" containerName="pruner" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.934999 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9bed013-a0a6-4038-9160-b24f47142f4f" containerName="pruner" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935059 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9bed013-a0a6-4038-9160-b24f47142f4f" containerName="pruner" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.935116 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935167 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.935225 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935281 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.935364 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935417 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.935554 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935616 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.935672 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935730 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.935799 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb6648af-1ab7-419e-bdb3-8f696aaabd80" containerName="oauth-openshift" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935855 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb6648af-1ab7-419e-bdb3-8f696aaabd80" containerName="oauth-openshift" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.935912 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.935971 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="extract-content" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.936024 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.936561 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.936629 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" containerName="collect-profiles" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.936686 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" containerName="collect-profiles" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.936740 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.936789 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.936875 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.936930 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.936981 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.937030 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="extract-utilities" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.937092 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.937149 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: E1124 19:18:37.937199 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.937406 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.937659 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2881c99f-084a-4c91-b4a6-8ebfddaaa4bb" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.937775 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="51cfcad3-1784-4f06-86bb-30612cf4308f" containerName="pruner" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.937852 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9bed013-a0a6-4038-9160-b24f47142f4f" containerName="pruner" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.938018 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e41759d-ff0c-412d-a9a6-7593b3ac69f0" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.938081 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6ccb2d-56f0-41a7-bc1b-5784948c3d28" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.938136 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="833de199-4de2-4674-8709-e15c8cd9b40c" containerName="registry-server" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.938201 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb6648af-1ab7-419e-bdb3-8f696aaabd80" containerName="oauth-openshift" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.938253 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" containerName="collect-profiles" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.938687 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:37 crc kubenswrapper[5035]: I1124 19:18:37.944921 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6644f974c8-r8xlb"] Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046051 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-provider-selection\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046123 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-dir\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046198 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-cliconfig\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046234 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-serving-cert\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046306 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-error\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046338 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhn2c\" (UniqueName: \"kubernetes.io/projected/fb6648af-1ab7-419e-bdb3-8f696aaabd80-kube-api-access-qhn2c\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046365 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-policies\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.046267 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.047432 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048128 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048264 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.047578 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-trusted-ca-bundle\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048450 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-ocp-branding-template\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048497 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-session\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048571 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-router-certs\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048619 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-service-ca\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048656 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-login\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.048707 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-idp-0-file-data\") pod \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\" (UID: \"fb6648af-1ab7-419e-bdb3-8f696aaabd80\") " Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049086 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-service-ca\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049134 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049166 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049192 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vnn6\" (UniqueName: \"kubernetes.io/projected/f1df3f95-6019-4e33-a827-f2d99ea28c4a-kube-api-access-5vnn6\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049284 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049342 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049446 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-error\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049499 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049537 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-login\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049569 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-session\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049672 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049707 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-audit-policies\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1df3f95-6019-4e33-a827-f2d99ea28c4a-audit-dir\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049875 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.049952 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-router-certs\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.050046 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.050069 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.050082 5035 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.050097 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.050111 5035 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fb6648af-1ab7-419e-bdb3-8f696aaabd80-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.053710 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.053776 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb6648af-1ab7-419e-bdb3-8f696aaabd80-kube-api-access-qhn2c" (OuterVolumeSpecName: "kube-api-access-qhn2c") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "kube-api-access-qhn2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.054123 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.054696 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.054995 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.055886 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.056075 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.056827 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.057267 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "fb6648af-1ab7-419e-bdb3-8f696aaabd80" (UID: "fb6648af-1ab7-419e-bdb3-8f696aaabd80"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.151441 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-service-ca\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.151971 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.152085 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.152188 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vnn6\" (UniqueName: \"kubernetes.io/projected/f1df3f95-6019-4e33-a827-f2d99ea28c4a-kube-api-access-5vnn6\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.152328 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.152443 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-error\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.152880 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.154266 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-login\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.155117 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-session\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.155375 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-audit-policies\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.155657 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.155885 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1df3f95-6019-4e33-a827-f2d99ea28c4a-audit-dir\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156077 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156237 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-service-ca\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.154085 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.155952 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.155976 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-error\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.155982 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1df3f95-6019-4e33-a827-f2d99ea28c4a-audit-dir\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.154182 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156178 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1df3f95-6019-4e33-a827-f2d99ea28c4a-audit-policies\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156258 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-router-certs\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156652 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156674 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156692 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156710 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156728 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156750 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhn2c\" (UniqueName: \"kubernetes.io/projected/fb6648af-1ab7-419e-bdb3-8f696aaabd80-kube-api-access-qhn2c\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156765 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156779 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.156796 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fb6648af-1ab7-419e-bdb3-8f696aaabd80-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.157572 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.158405 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-session\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.159603 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-user-template-login\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.160177 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.160554 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.162203 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f1df3f95-6019-4e33-a827-f2d99ea28c4a-v4-0-config-system-router-certs\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.172583 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vnn6\" (UniqueName: \"kubernetes.io/projected/f1df3f95-6019-4e33-a827-f2d99ea28c4a-kube-api-access-5vnn6\") pod \"oauth-openshift-6644f974c8-r8xlb\" (UID: \"f1df3f95-6019-4e33-a827-f2d99ea28c4a\") " pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.261794 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.520489 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6644f974c8-r8xlb"] Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.575376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" event={"ID":"f1df3f95-6019-4e33-a827-f2d99ea28c4a","Type":"ContainerStarted","Data":"8f7becb88cca9438cc893af0430edc83c9e7fa98b6f4565d6956275f58a44a28"} Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.576926 5035 generic.go:334] "Generic (PLEG): container finished" podID="fb6648af-1ab7-419e-bdb3-8f696aaabd80" containerID="6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de" exitCode=0 Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.576974 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" event={"ID":"fb6648af-1ab7-419e-bdb3-8f696aaabd80","Type":"ContainerDied","Data":"6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de"} Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.576985 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.577008 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qfxzs" event={"ID":"fb6648af-1ab7-419e-bdb3-8f696aaabd80","Type":"ContainerDied","Data":"c49f83fa2044b3073d693302b85d5bc2a8eb21f5220d778554c70791bee18b06"} Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.577029 5035 scope.go:117] "RemoveContainer" containerID="6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.603517 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qfxzs"] Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.604558 5035 scope.go:117] "RemoveContainer" containerID="6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de" Nov 24 19:18:38 crc kubenswrapper[5035]: E1124 19:18:38.605075 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de\": container with ID starting with 6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de not found: ID does not exist" containerID="6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.605122 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de"} err="failed to get container status \"6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de\": rpc error: code = NotFound desc = could not find container \"6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de\": container with ID starting with 6a44e98cb0bde03c8a3b4a89cef4a46dcf8b5936eead6d5d6c358448852f07de not found: ID does not exist" Nov 24 19:18:38 crc kubenswrapper[5035]: I1124 19:18:38.607646 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qfxzs"] Nov 24 19:18:39 crc kubenswrapper[5035]: I1124 19:18:39.591024 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" event={"ID":"f1df3f95-6019-4e33-a827-f2d99ea28c4a","Type":"ContainerStarted","Data":"dde1fff7a9cfdeae37f4946c20800c30c11210128a541a6b4558a5c97fb03d8f"} Nov 24 19:18:39 crc kubenswrapper[5035]: I1124 19:18:39.592383 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:39 crc kubenswrapper[5035]: I1124 19:18:39.600564 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" Nov 24 19:18:39 crc kubenswrapper[5035]: I1124 19:18:39.615363 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6644f974c8-r8xlb" podStartSLOduration=27.615333834 podStartE2EDuration="27.615333834s" podCreationTimestamp="2025-11-24 19:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:18:39.610075131 +0000 UTC m=+258.132581408" watchObservedRunningTime="2025-11-24 19:18:39.615333834 +0000 UTC m=+258.137840111" Nov 24 19:18:40 crc kubenswrapper[5035]: I1124 19:18:40.206339 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb6648af-1ab7-419e-bdb3-8f696aaabd80" path="/var/lib/kubelet/pods/fb6648af-1ab7-419e-bdb3-8f696aaabd80/volumes" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.228254 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8sczb"] Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.229048 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8sczb" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="registry-server" containerID="cri-o://17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd" gracePeriod=30 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.244505 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlgkh"] Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.244773 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dlgkh" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="registry-server" containerID="cri-o://d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" gracePeriod=30 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.251321 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q2jhw"] Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.251551 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerName="marketplace-operator" containerID="cri-o://e116fe9499a7fe2502e95703f1a7252a24d2393944938a45c46e89aa2451ce8d" gracePeriod=30 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.261827 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-swgf2"] Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.262119 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-swgf2" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="registry-server" containerID="cri-o://dc525aa8e4bf2a4f4889de7d87e48b7da78f13f3a232640cb7cfe3556a214c67" gracePeriod=30 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.265601 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49zbg"] Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.266420 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.269373 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5ltvz"] Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.269662 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5ltvz" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="registry-server" containerID="cri-o://fdd1ffb8521e97699a8b5d667d66fe5f7824eb8404650bc74c712d1e265f610d" gracePeriod=30 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.288665 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49zbg"] Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.394473 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3 is running failed: container process not found" containerID="d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.394912 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3 is running failed: container process not found" containerID="d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.395372 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3 is running failed: container process not found" containerID="d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" cmd=["grpc_health_probe","-addr=:50051"] Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.395496 5035 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-dlgkh" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="registry-server" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.436879 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89zzx\" (UniqueName: \"kubernetes.io/projected/70cbf730-06f3-4b21-aa8c-64a80ae4929b-kube-api-access-89zzx\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.436942 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70cbf730-06f3-4b21-aa8c-64a80ae4929b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.436964 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70cbf730-06f3-4b21-aa8c-64a80ae4929b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.542654 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89zzx\" (UniqueName: \"kubernetes.io/projected/70cbf730-06f3-4b21-aa8c-64a80ae4929b-kube-api-access-89zzx\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.542709 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70cbf730-06f3-4b21-aa8c-64a80ae4929b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.542731 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70cbf730-06f3-4b21-aa8c-64a80ae4929b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.545229 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70cbf730-06f3-4b21-aa8c-64a80ae4929b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.560657 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89zzx\" (UniqueName: \"kubernetes.io/projected/70cbf730-06f3-4b21-aa8c-64a80ae4929b-kube-api-access-89zzx\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.561396 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70cbf730-06f3-4b21-aa8c-64a80ae4929b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49zbg\" (UID: \"70cbf730-06f3-4b21-aa8c-64a80ae4929b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.738375 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.744920 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.749018 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.811041 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8sczb" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.811032 5035 generic.go:334] "Generic (PLEG): container finished" podID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerID="17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd" exitCode=0 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.811248 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8sczb" event={"ID":"0d029a23-831f-48f1-939e-de845c7a7c5b","Type":"ContainerDied","Data":"17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd"} Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.811310 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8sczb" event={"ID":"0d029a23-831f-48f1-939e-de845c7a7c5b","Type":"ContainerDied","Data":"528b81faf6b89f71620107de9c4d7ff943d9ae8bd7e047c0dfc56eb9c12d9c81"} Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.811335 5035 scope.go:117] "RemoveContainer" containerID="17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.834374 5035 generic.go:334] "Generic (PLEG): container finished" podID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerID="d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" exitCode=0 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.834451 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlgkh" event={"ID":"bc56ef71-4ba1-4378-b7ea-a97141be037e","Type":"ContainerDied","Data":"d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3"} Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.834480 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlgkh" event={"ID":"bc56ef71-4ba1-4378-b7ea-a97141be037e","Type":"ContainerDied","Data":"e3322c80489bdb4e309d06b6713dc1c7ce9f39ef23811f1527c75ae3fe4b29af"} Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.834554 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlgkh" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.838234 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.840490 5035 generic.go:334] "Generic (PLEG): container finished" podID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerID="fdd1ffb8521e97699a8b5d667d66fe5f7824eb8404650bc74c712d1e265f610d" exitCode=0 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.840659 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ltvz" event={"ID":"5efc44e8-4d11-4f01-b819-e56df5f4d88a","Type":"ContainerDied","Data":"fdd1ffb8521e97699a8b5d667d66fe5f7824eb8404650bc74c712d1e265f610d"} Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.844046 5035 scope.go:117] "RemoveContainer" containerID="ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.845728 5035 generic.go:334] "Generic (PLEG): container finished" podID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerID="e116fe9499a7fe2502e95703f1a7252a24d2393944938a45c46e89aa2451ce8d" exitCode=0 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.845818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" event={"ID":"fa706698-c956-4fe2-aaad-e4feb290d36e","Type":"ContainerDied","Data":"e116fe9499a7fe2502e95703f1a7252a24d2393944938a45c46e89aa2451ce8d"} Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.864306 5035 generic.go:334] "Generic (PLEG): container finished" podID="7c332553-06e8-4b00-9c70-0feffd93f281" containerID="dc525aa8e4bf2a4f4889de7d87e48b7da78f13f3a232640cb7cfe3556a214c67" exitCode=0 Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.864354 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swgf2" event={"ID":"7c332553-06e8-4b00-9c70-0feffd93f281","Type":"ContainerDied","Data":"dc525aa8e4bf2a4f4889de7d87e48b7da78f13f3a232640cb7cfe3556a214c67"} Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.866530 5035 scope.go:117] "RemoveContainer" containerID="39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.887410 5035 scope.go:117] "RemoveContainer" containerID="17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd" Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.887844 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd\": container with ID starting with 17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd not found: ID does not exist" containerID="17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.887886 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd"} err="failed to get container status \"17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd\": rpc error: code = NotFound desc = could not find container \"17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd\": container with ID starting with 17f817a717426ee9b8660b37df2f12acbc2017daee49bf60ce2b8c9fd67d37dd not found: ID does not exist" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.887908 5035 scope.go:117] "RemoveContainer" containerID="ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1" Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.888176 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1\": container with ID starting with ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1 not found: ID does not exist" containerID="ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.888201 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1"} err="failed to get container status \"ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1\": rpc error: code = NotFound desc = could not find container \"ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1\": container with ID starting with ea56ad6a3bc1ab00903166977c4ba8a1971de919424c469d249bb50697f180e1 not found: ID does not exist" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.888214 5035 scope.go:117] "RemoveContainer" containerID="39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188" Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.888429 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188\": container with ID starting with 39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188 not found: ID does not exist" containerID="39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.888453 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188"} err="failed to get container status \"39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188\": rpc error: code = NotFound desc = could not find container \"39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188\": container with ID starting with 39d9f4918876605fff903bcca77ebafc96967f5b46a31624e32f764c8028b188 not found: ID does not exist" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.888466 5035 scope.go:117] "RemoveContainer" containerID="d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.906338 5035 scope.go:117] "RemoveContainer" containerID="aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.916499 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.923278 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.930128 5035 scope.go:117] "RemoveContainer" containerID="36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-catalog-content\") pod \"7c332553-06e8-4b00-9c70-0feffd93f281\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949223 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-operator-metrics\") pod \"fa706698-c956-4fe2-aaad-e4feb290d36e\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949246 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-utilities\") pod \"7c332553-06e8-4b00-9c70-0feffd93f281\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949274 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-utilities\") pod \"0d029a23-831f-48f1-939e-de845c7a7c5b\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949312 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jklxv\" (UniqueName: \"kubernetes.io/projected/bc56ef71-4ba1-4378-b7ea-a97141be037e-kube-api-access-jklxv\") pod \"bc56ef71-4ba1-4378-b7ea-a97141be037e\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949333 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf8zh\" (UniqueName: \"kubernetes.io/projected/0d029a23-831f-48f1-939e-de845c7a7c5b-kube-api-access-sf8zh\") pod \"0d029a23-831f-48f1-939e-de845c7a7c5b\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949357 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vdwm\" (UniqueName: \"kubernetes.io/projected/fa706698-c956-4fe2-aaad-e4feb290d36e-kube-api-access-6vdwm\") pod \"fa706698-c956-4fe2-aaad-e4feb290d36e\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949378 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzxrn\" (UniqueName: \"kubernetes.io/projected/7c332553-06e8-4b00-9c70-0feffd93f281-kube-api-access-wzxrn\") pod \"7c332553-06e8-4b00-9c70-0feffd93f281\" (UID: \"7c332553-06e8-4b00-9c70-0feffd93f281\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949405 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-trusted-ca\") pod \"fa706698-c956-4fe2-aaad-e4feb290d36e\" (UID: \"fa706698-c956-4fe2-aaad-e4feb290d36e\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949427 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-utilities\") pod \"bc56ef71-4ba1-4378-b7ea-a97141be037e\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949446 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-catalog-content\") pod \"bc56ef71-4ba1-4378-b7ea-a97141be037e\" (UID: \"bc56ef71-4ba1-4378-b7ea-a97141be037e\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949468 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27kfq\" (UniqueName: \"kubernetes.io/projected/5efc44e8-4d11-4f01-b819-e56df5f4d88a-kube-api-access-27kfq\") pod \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949504 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-utilities\") pod \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949529 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-catalog-content\") pod \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\" (UID: \"5efc44e8-4d11-4f01-b819-e56df5f4d88a\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.949569 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-catalog-content\") pod \"0d029a23-831f-48f1-939e-de845c7a7c5b\" (UID: \"0d029a23-831f-48f1-939e-de845c7a7c5b\") " Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.956412 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa706698-c956-4fe2-aaad-e4feb290d36e-kube-api-access-6vdwm" (OuterVolumeSpecName: "kube-api-access-6vdwm") pod "fa706698-c956-4fe2-aaad-e4feb290d36e" (UID: "fa706698-c956-4fe2-aaad-e4feb290d36e"). InnerVolumeSpecName "kube-api-access-6vdwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.964874 5035 scope.go:117] "RemoveContainer" containerID="d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.965098 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-utilities" (OuterVolumeSpecName: "utilities") pod "0d029a23-831f-48f1-939e-de845c7a7c5b" (UID: "0d029a23-831f-48f1-939e-de845c7a7c5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.965736 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-utilities" (OuterVolumeSpecName: "utilities") pod "7c332553-06e8-4b00-9c70-0feffd93f281" (UID: "7c332553-06e8-4b00-9c70-0feffd93f281"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.967412 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3\": container with ID starting with d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3 not found: ID does not exist" containerID="d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.967448 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3"} err="failed to get container status \"d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3\": rpc error: code = NotFound desc = could not find container \"d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3\": container with ID starting with d6df4d7e6db68bf7e9c41284b4efd9df649d7ef68a711ac0f3b8bfb231f5fee3 not found: ID does not exist" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.967470 5035 scope.go:117] "RemoveContainer" containerID="aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.968108 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-utilities" (OuterVolumeSpecName: "utilities") pod "5efc44e8-4d11-4f01-b819-e56df5f4d88a" (UID: "5efc44e8-4d11-4f01-b819-e56df5f4d88a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.968547 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "fa706698-c956-4fe2-aaad-e4feb290d36e" (UID: "fa706698-c956-4fe2-aaad-e4feb290d36e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.969608 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8\": container with ID starting with aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8 not found: ID does not exist" containerID="aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.969640 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8"} err="failed to get container status \"aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8\": rpc error: code = NotFound desc = could not find container \"aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8\": container with ID starting with aa8cf35dc21da15ce2af14a88f63d02ee1d0b19e444217f3c7abd3d06dc4a7d8 not found: ID does not exist" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.969662 5035 scope.go:117] "RemoveContainer" containerID="36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f" Nov 24 19:19:00 crc kubenswrapper[5035]: E1124 19:19:00.970579 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f\": container with ID starting with 36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f not found: ID does not exist" containerID="36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.971280 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-utilities" (OuterVolumeSpecName: "utilities") pod "bc56ef71-4ba1-4378-b7ea-a97141be037e" (UID: "bc56ef71-4ba1-4378-b7ea-a97141be037e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.970659 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f"} err="failed to get container status \"36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f\": rpc error: code = NotFound desc = could not find container \"36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f\": container with ID starting with 36cdb8336c32a26e9279b0962ea3c86ed7d9cb72bb372fe478d24d8150b8141f not found: ID does not exist" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.972230 5035 scope.go:117] "RemoveContainer" containerID="fdd1ffb8521e97699a8b5d667d66fe5f7824eb8404650bc74c712d1e265f610d" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.973483 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c332553-06e8-4b00-9c70-0feffd93f281-kube-api-access-wzxrn" (OuterVolumeSpecName: "kube-api-access-wzxrn") pod "7c332553-06e8-4b00-9c70-0feffd93f281" (UID: "7c332553-06e8-4b00-9c70-0feffd93f281"). InnerVolumeSpecName "kube-api-access-wzxrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.975523 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5efc44e8-4d11-4f01-b819-e56df5f4d88a-kube-api-access-27kfq" (OuterVolumeSpecName: "kube-api-access-27kfq") pod "5efc44e8-4d11-4f01-b819-e56df5f4d88a" (UID: "5efc44e8-4d11-4f01-b819-e56df5f4d88a"). InnerVolumeSpecName "kube-api-access-27kfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.982374 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "fa706698-c956-4fe2-aaad-e4feb290d36e" (UID: "fa706698-c956-4fe2-aaad-e4feb290d36e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.979419 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d029a23-831f-48f1-939e-de845c7a7c5b-kube-api-access-sf8zh" (OuterVolumeSpecName: "kube-api-access-sf8zh") pod "0d029a23-831f-48f1-939e-de845c7a7c5b" (UID: "0d029a23-831f-48f1-939e-de845c7a7c5b"). InnerVolumeSpecName "kube-api-access-sf8zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:19:00 crc kubenswrapper[5035]: I1124 19:19:00.996875 5035 scope.go:117] "RemoveContainer" containerID="78af914582f6e3c310f4d8f84ceeb54ad798c4758f3ffc5aa7525c0a248ce8e1" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.004557 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc56ef71-4ba1-4378-b7ea-a97141be037e-kube-api-access-jklxv" (OuterVolumeSpecName: "kube-api-access-jklxv") pod "bc56ef71-4ba1-4378-b7ea-a97141be037e" (UID: "bc56ef71-4ba1-4378-b7ea-a97141be037e"). InnerVolumeSpecName "kube-api-access-jklxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.017989 5035 scope.go:117] "RemoveContainer" containerID="24d2e74d00b063c17fae6933934961e3c96a01eb687fce58bc22ddf6b38e186a" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.022819 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c332553-06e8-4b00-9c70-0feffd93f281" (UID: "7c332553-06e8-4b00-9c70-0feffd93f281"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.037135 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49zbg"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.050587 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.050754 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.050839 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c332553-06e8-4b00-9c70-0feffd93f281-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.050946 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051038 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jklxv\" (UniqueName: \"kubernetes.io/projected/bc56ef71-4ba1-4378-b7ea-a97141be037e-kube-api-access-jklxv\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051104 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf8zh\" (UniqueName: \"kubernetes.io/projected/0d029a23-831f-48f1-939e-de845c7a7c5b-kube-api-access-sf8zh\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051159 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vdwm\" (UniqueName: \"kubernetes.io/projected/fa706698-c956-4fe2-aaad-e4feb290d36e-kube-api-access-6vdwm\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051222 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzxrn\" (UniqueName: \"kubernetes.io/projected/7c332553-06e8-4b00-9c70-0feffd93f281-kube-api-access-wzxrn\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051283 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa706698-c956-4fe2-aaad-e4feb290d36e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051370 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051429 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27kfq\" (UniqueName: \"kubernetes.io/projected/5efc44e8-4d11-4f01-b819-e56df5f4d88a-kube-api-access-27kfq\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.051483 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.068905 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc56ef71-4ba1-4378-b7ea-a97141be037e" (UID: "bc56ef71-4ba1-4378-b7ea-a97141be037e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.078319 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d029a23-831f-48f1-939e-de845c7a7c5b" (UID: "0d029a23-831f-48f1-939e-de845c7a7c5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.088923 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5efc44e8-4d11-4f01-b819-e56df5f4d88a" (UID: "5efc44e8-4d11-4f01-b819-e56df5f4d88a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.138153 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8sczb"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.140458 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8sczb"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.153259 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d029a23-831f-48f1-939e-de845c7a7c5b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.153324 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc56ef71-4ba1-4378-b7ea-a97141be037e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.153338 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5efc44e8-4d11-4f01-b819-e56df5f4d88a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.166886 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlgkh"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.169973 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dlgkh"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.870491 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ltvz" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.870477 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ltvz" event={"ID":"5efc44e8-4d11-4f01-b819-e56df5f4d88a","Type":"ContainerDied","Data":"68435385d934664ad6fc977dbc28f1dc5230412b5fd5d67c348c109a7666b120"} Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.871876 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" event={"ID":"70cbf730-06f3-4b21-aa8c-64a80ae4929b","Type":"ContainerStarted","Data":"a9e171a82bd5ac773f0eb0e0ba58202c9842090c3240e1884a843661452c3404"} Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.871946 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.871968 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" event={"ID":"70cbf730-06f3-4b21-aa8c-64a80ae4929b","Type":"ContainerStarted","Data":"90052e6a972fd406104da85412de1a9856d574467fd1f4e3b4562083873ded0e"} Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.873637 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" event={"ID":"fa706698-c956-4fe2-aaad-e4feb290d36e","Type":"ContainerDied","Data":"40208ea8593284174e68f425948d633557e6c419e14f135ea7f9673c329deeee"} Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.873685 5035 scope.go:117] "RemoveContainer" containerID="e116fe9499a7fe2502e95703f1a7252a24d2393944938a45c46e89aa2451ce8d" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.873789 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-q2jhw" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.876919 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swgf2" event={"ID":"7c332553-06e8-4b00-9c70-0feffd93f281","Type":"ContainerDied","Data":"f58c8ba11a756d6439a809819fd64dd3cce4ff1641e9ce1a314921a903b01128"} Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.876971 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swgf2" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.877229 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.905109 5035 scope.go:117] "RemoveContainer" containerID="dc525aa8e4bf2a4f4889de7d87e48b7da78f13f3a232640cb7cfe3556a214c67" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.908664 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-49zbg" podStartSLOduration=1.9086507990000001 podStartE2EDuration="1.908650799s" podCreationTimestamp="2025-11-24 19:19:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:19:01.894566826 +0000 UTC m=+280.417073163" watchObservedRunningTime="2025-11-24 19:19:01.908650799 +0000 UTC m=+280.431157056" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.912739 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5ltvz"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.919791 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5ltvz"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.921248 5035 scope.go:117] "RemoveContainer" containerID="5df1613c5757f61dd55605151cc1a3872380dff9977d497cbb36c3389c91fc50" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.927411 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-swgf2"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.932452 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-swgf2"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.946027 5035 scope.go:117] "RemoveContainer" containerID="e37d4605c669733317a1c3ff682c102723e1a4c7a1188efae2904eff5e154307" Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.962761 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q2jhw"] Nov 24 19:19:01 crc kubenswrapper[5035]: I1124 19:19:01.964774 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-q2jhw"] Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041145 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k54vj"] Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041383 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041404 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041416 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041424 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041436 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041443 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041456 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041464 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041477 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerName="marketplace-operator" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041485 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerName="marketplace-operator" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041492 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041502 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041513 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041522 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041532 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041542 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041555 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041562 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041573 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041580 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="extract-utilities" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041591 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041598 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041609 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041616 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="extract-content" Nov 24 19:19:02 crc kubenswrapper[5035]: E1124 19:19:02.041627 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041634 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041723 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041738 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041750 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041761 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" containerName="registry-server" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.041769 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" containerName="marketplace-operator" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.042681 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.046188 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.054093 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k54vj"] Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.062100 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-utilities\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.062149 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fq8h\" (UniqueName: \"kubernetes.io/projected/2295440b-2755-4332-8d15-5e2afdeefc9e-kube-api-access-6fq8h\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.062182 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-catalog-content\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.163593 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-catalog-content\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.163687 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-utilities\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.163720 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fq8h\" (UniqueName: \"kubernetes.io/projected/2295440b-2755-4332-8d15-5e2afdeefc9e-kube-api-access-6fq8h\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.164071 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-catalog-content\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.164313 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-utilities\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.183232 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fq8h\" (UniqueName: \"kubernetes.io/projected/2295440b-2755-4332-8d15-5e2afdeefc9e-kube-api-access-6fq8h\") pod \"certified-operators-k54vj\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.207067 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d029a23-831f-48f1-939e-de845c7a7c5b" path="/var/lib/kubelet/pods/0d029a23-831f-48f1-939e-de845c7a7c5b/volumes" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.207864 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5efc44e8-4d11-4f01-b819-e56df5f4d88a" path="/var/lib/kubelet/pods/5efc44e8-4d11-4f01-b819-e56df5f4d88a/volumes" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.208584 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c332553-06e8-4b00-9c70-0feffd93f281" path="/var/lib/kubelet/pods/7c332553-06e8-4b00-9c70-0feffd93f281/volumes" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.209764 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc56ef71-4ba1-4378-b7ea-a97141be037e" path="/var/lib/kubelet/pods/bc56ef71-4ba1-4378-b7ea-a97141be037e/volumes" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.211085 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa706698-c956-4fe2-aaad-e4feb290d36e" path="/var/lib/kubelet/pods/fa706698-c956-4fe2-aaad-e4feb290d36e/volumes" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.356384 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.528408 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k54vj"] Nov 24 19:19:02 crc kubenswrapper[5035]: W1124 19:19:02.535850 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2295440b_2755_4332_8d15_5e2afdeefc9e.slice/crio-7f7684801806a4c8b3c59e431012d33d8547944b46089c64de6ca0d6433581b4 WatchSource:0}: Error finding container 7f7684801806a4c8b3c59e431012d33d8547944b46089c64de6ca0d6433581b4: Status 404 returned error can't find the container with id 7f7684801806a4c8b3c59e431012d33d8547944b46089c64de6ca0d6433581b4 Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.887575 5035 generic.go:334] "Generic (PLEG): container finished" podID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerID="5d9dd9e9903da1ec31180ce4be78fc8f2a2d1a39f733518951548663af1dc237" exitCode=0 Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.887690 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k54vj" event={"ID":"2295440b-2755-4332-8d15-5e2afdeefc9e","Type":"ContainerDied","Data":"5d9dd9e9903da1ec31180ce4be78fc8f2a2d1a39f733518951548663af1dc237"} Nov 24 19:19:02 crc kubenswrapper[5035]: I1124 19:19:02.887735 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k54vj" event={"ID":"2295440b-2755-4332-8d15-5e2afdeefc9e","Type":"ContainerStarted","Data":"7f7684801806a4c8b3c59e431012d33d8547944b46089c64de6ca0d6433581b4"} Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.046369 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x2vrj"] Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.054224 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.059812 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.060564 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x2vrj"] Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.075801 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbj6j\" (UniqueName: \"kubernetes.io/projected/e5328a7d-e2f7-43ae-89b9-c264960c8912-kube-api-access-hbj6j\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.075849 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5328a7d-e2f7-43ae-89b9-c264960c8912-utilities\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.075868 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5328a7d-e2f7-43ae-89b9-c264960c8912-catalog-content\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.177677 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbj6j\" (UniqueName: \"kubernetes.io/projected/e5328a7d-e2f7-43ae-89b9-c264960c8912-kube-api-access-hbj6j\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.178091 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5328a7d-e2f7-43ae-89b9-c264960c8912-utilities\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.178406 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5328a7d-e2f7-43ae-89b9-c264960c8912-catalog-content\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.178678 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5328a7d-e2f7-43ae-89b9-c264960c8912-utilities\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.178811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5328a7d-e2f7-43ae-89b9-c264960c8912-catalog-content\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.204084 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbj6j\" (UniqueName: \"kubernetes.io/projected/e5328a7d-e2f7-43ae-89b9-c264960c8912-kube-api-access-hbj6j\") pod \"redhat-marketplace-x2vrj\" (UID: \"e5328a7d-e2f7-43ae-89b9-c264960c8912\") " pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.414989 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.810784 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x2vrj"] Nov 24 19:19:03 crc kubenswrapper[5035]: I1124 19:19:03.894889 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x2vrj" event={"ID":"e5328a7d-e2f7-43ae-89b9-c264960c8912","Type":"ContainerStarted","Data":"4c20091df8f4f82ada0f93f45ecd14dd47e15829049cfc9cd8bdc41c6a845a81"} Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.450052 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s9kng"] Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.451035 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.458016 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.463903 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9kng"] Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.494883 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-catalog-content\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.494989 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpnt6\" (UniqueName: \"kubernetes.io/projected/62284293-00ac-4dd5-8e14-ecebe023008c-kube-api-access-kpnt6\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.495030 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-utilities\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.596222 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-catalog-content\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.596733 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-catalog-content\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.596874 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpnt6\" (UniqueName: \"kubernetes.io/projected/62284293-00ac-4dd5-8e14-ecebe023008c-kube-api-access-kpnt6\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.596905 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-utilities\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.597242 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-utilities\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.617579 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpnt6\" (UniqueName: \"kubernetes.io/projected/62284293-00ac-4dd5-8e14-ecebe023008c-kube-api-access-kpnt6\") pod \"redhat-operators-s9kng\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.784489 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.902848 5035 generic.go:334] "Generic (PLEG): container finished" podID="e5328a7d-e2f7-43ae-89b9-c264960c8912" containerID="28c8e9f0cc1f0582173c8017904d880ff761e472d951123a30f165827bcebab7" exitCode=0 Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.902907 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x2vrj" event={"ID":"e5328a7d-e2f7-43ae-89b9-c264960c8912","Type":"ContainerDied","Data":"28c8e9f0cc1f0582173c8017904d880ff761e472d951123a30f165827bcebab7"} Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.906437 5035 generic.go:334] "Generic (PLEG): container finished" podID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerID="52023767e730293ce283ce003facee4708e99e2a3a5bf21d941e9bf7ef9d2036" exitCode=0 Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.906468 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k54vj" event={"ID":"2295440b-2755-4332-8d15-5e2afdeefc9e","Type":"ContainerDied","Data":"52023767e730293ce283ce003facee4708e99e2a3a5bf21d941e9bf7ef9d2036"} Nov 24 19:19:04 crc kubenswrapper[5035]: I1124 19:19:04.988468 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s9kng"] Nov 24 19:19:04 crc kubenswrapper[5035]: W1124 19:19:04.999311 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62284293_00ac_4dd5_8e14_ecebe023008c.slice/crio-161a013d6fbd24dbdd60ebaca4e098b641e3c4b1b957f1951d2cc32acf2b9479 WatchSource:0}: Error finding container 161a013d6fbd24dbdd60ebaca4e098b641e3c4b1b957f1951d2cc32acf2b9479: Status 404 returned error can't find the container with id 161a013d6fbd24dbdd60ebaca4e098b641e3c4b1b957f1951d2cc32acf2b9479 Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.450327 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tr6hr"] Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.451879 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.453895 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.459131 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tr6hr"] Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.608335 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-utilities\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.608384 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-catalog-content\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.608446 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnfhz\" (UniqueName: \"kubernetes.io/projected/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-kube-api-access-pnfhz\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.714575 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-utilities\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.714630 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-catalog-content\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.714692 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnfhz\" (UniqueName: \"kubernetes.io/projected/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-kube-api-access-pnfhz\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.715163 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-utilities\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.715327 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-catalog-content\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.738115 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnfhz\" (UniqueName: \"kubernetes.io/projected/14a39f18-63d0-4f2d-a1d8-e8117872b9f6-kube-api-access-pnfhz\") pod \"community-operators-tr6hr\" (UID: \"14a39f18-63d0-4f2d-a1d8-e8117872b9f6\") " pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.783604 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.931553 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k54vj" event={"ID":"2295440b-2755-4332-8d15-5e2afdeefc9e","Type":"ContainerStarted","Data":"3a3fdb05890dc52f33afdd871678f92398a79d895e3fba8552600c255db7540f"} Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.940412 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x2vrj" event={"ID":"e5328a7d-e2f7-43ae-89b9-c264960c8912","Type":"ContainerStarted","Data":"fc585597a7043ec4ffb4f27f973c6205fd3586ce39410803789304f6421dc138"} Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.944939 5035 generic.go:334] "Generic (PLEG): container finished" podID="62284293-00ac-4dd5-8e14-ecebe023008c" containerID="b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12" exitCode=0 Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.944986 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9kng" event={"ID":"62284293-00ac-4dd5-8e14-ecebe023008c","Type":"ContainerDied","Data":"b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12"} Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.945014 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9kng" event={"ID":"62284293-00ac-4dd5-8e14-ecebe023008c","Type":"ContainerStarted","Data":"161a013d6fbd24dbdd60ebaca4e098b641e3c4b1b957f1951d2cc32acf2b9479"} Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.951944 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k54vj" podStartSLOduration=1.448485203 podStartE2EDuration="3.951924492s" podCreationTimestamp="2025-11-24 19:19:02 +0000 UTC" firstStartedPulling="2025-11-24 19:19:02.889046409 +0000 UTC m=+281.411552666" lastFinishedPulling="2025-11-24 19:19:05.392485698 +0000 UTC m=+283.914991955" observedRunningTime="2025-11-24 19:19:05.948422016 +0000 UTC m=+284.470928273" watchObservedRunningTime="2025-11-24 19:19:05.951924492 +0000 UTC m=+284.474430749" Nov 24 19:19:05 crc kubenswrapper[5035]: I1124 19:19:05.989902 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tr6hr"] Nov 24 19:19:05 crc kubenswrapper[5035]: W1124 19:19:05.994052 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14a39f18_63d0_4f2d_a1d8_e8117872b9f6.slice/crio-0b15918153ac55f1c2fea8608a391c512c959d344ba2771dd6ce51ead18b3ebc WatchSource:0}: Error finding container 0b15918153ac55f1c2fea8608a391c512c959d344ba2771dd6ce51ead18b3ebc: Status 404 returned error can't find the container with id 0b15918153ac55f1c2fea8608a391c512c959d344ba2771dd6ce51ead18b3ebc Nov 24 19:19:06 crc kubenswrapper[5035]: I1124 19:19:06.952027 5035 generic.go:334] "Generic (PLEG): container finished" podID="14a39f18-63d0-4f2d-a1d8-e8117872b9f6" containerID="e0a43d304e41e23507c5eb48e4ce37a1bb298a457c9b344e05b2649158e2a69c" exitCode=0 Nov 24 19:19:06 crc kubenswrapper[5035]: I1124 19:19:06.952107 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr6hr" event={"ID":"14a39f18-63d0-4f2d-a1d8-e8117872b9f6","Type":"ContainerDied","Data":"e0a43d304e41e23507c5eb48e4ce37a1bb298a457c9b344e05b2649158e2a69c"} Nov 24 19:19:06 crc kubenswrapper[5035]: I1124 19:19:06.952650 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr6hr" event={"ID":"14a39f18-63d0-4f2d-a1d8-e8117872b9f6","Type":"ContainerStarted","Data":"0b15918153ac55f1c2fea8608a391c512c959d344ba2771dd6ce51ead18b3ebc"} Nov 24 19:19:06 crc kubenswrapper[5035]: I1124 19:19:06.960007 5035 generic.go:334] "Generic (PLEG): container finished" podID="e5328a7d-e2f7-43ae-89b9-c264960c8912" containerID="fc585597a7043ec4ffb4f27f973c6205fd3586ce39410803789304f6421dc138" exitCode=0 Nov 24 19:19:06 crc kubenswrapper[5035]: I1124 19:19:06.960811 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x2vrj" event={"ID":"e5328a7d-e2f7-43ae-89b9-c264960c8912","Type":"ContainerDied","Data":"fc585597a7043ec4ffb4f27f973c6205fd3586ce39410803789304f6421dc138"} Nov 24 19:19:07 crc kubenswrapper[5035]: I1124 19:19:07.967155 5035 generic.go:334] "Generic (PLEG): container finished" podID="62284293-00ac-4dd5-8e14-ecebe023008c" containerID="2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b" exitCode=0 Nov 24 19:19:07 crc kubenswrapper[5035]: I1124 19:19:07.967214 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9kng" event={"ID":"62284293-00ac-4dd5-8e14-ecebe023008c","Type":"ContainerDied","Data":"2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b"} Nov 24 19:19:07 crc kubenswrapper[5035]: I1124 19:19:07.972096 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr6hr" event={"ID":"14a39f18-63d0-4f2d-a1d8-e8117872b9f6","Type":"ContainerStarted","Data":"aed2e47c8ba6d6e4a7a024e2e5152fd58a2efd32ecfc60ae0f02670c67af68be"} Nov 24 19:19:07 crc kubenswrapper[5035]: I1124 19:19:07.974930 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x2vrj" event={"ID":"e5328a7d-e2f7-43ae-89b9-c264960c8912","Type":"ContainerStarted","Data":"f4e08b0df83feb4777f40e91183f0d3727a56a1b98e7ae49fb703716f205e7a7"} Nov 24 19:19:08 crc kubenswrapper[5035]: I1124 19:19:08.005325 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x2vrj" podStartSLOduration=2.529836853 podStartE2EDuration="5.00528078s" podCreationTimestamp="2025-11-24 19:19:03 +0000 UTC" firstStartedPulling="2025-11-24 19:19:04.906083371 +0000 UTC m=+283.428589628" lastFinishedPulling="2025-11-24 19:19:07.381527298 +0000 UTC m=+285.904033555" observedRunningTime="2025-11-24 19:19:08.004249642 +0000 UTC m=+286.526755899" watchObservedRunningTime="2025-11-24 19:19:08.00528078 +0000 UTC m=+286.527787037" Nov 24 19:19:08 crc kubenswrapper[5035]: I1124 19:19:08.982423 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9kng" event={"ID":"62284293-00ac-4dd5-8e14-ecebe023008c","Type":"ContainerStarted","Data":"e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5"} Nov 24 19:19:08 crc kubenswrapper[5035]: I1124 19:19:08.984886 5035 generic.go:334] "Generic (PLEG): container finished" podID="14a39f18-63d0-4f2d-a1d8-e8117872b9f6" containerID="aed2e47c8ba6d6e4a7a024e2e5152fd58a2efd32ecfc60ae0f02670c67af68be" exitCode=0 Nov 24 19:19:08 crc kubenswrapper[5035]: I1124 19:19:08.985610 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr6hr" event={"ID":"14a39f18-63d0-4f2d-a1d8-e8117872b9f6","Type":"ContainerDied","Data":"aed2e47c8ba6d6e4a7a024e2e5152fd58a2efd32ecfc60ae0f02670c67af68be"} Nov 24 19:19:09 crc kubenswrapper[5035]: I1124 19:19:09.010076 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s9kng" podStartSLOduration=2.547273301 podStartE2EDuration="5.010055244s" podCreationTimestamp="2025-11-24 19:19:04 +0000 UTC" firstStartedPulling="2025-11-24 19:19:05.946782522 +0000 UTC m=+284.469288779" lastFinishedPulling="2025-11-24 19:19:08.409564465 +0000 UTC m=+286.932070722" observedRunningTime="2025-11-24 19:19:09.00514403 +0000 UTC m=+287.527650287" watchObservedRunningTime="2025-11-24 19:19:09.010055244 +0000 UTC m=+287.532561501" Nov 24 19:19:11 crc kubenswrapper[5035]: I1124 19:19:11.006474 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr6hr" event={"ID":"14a39f18-63d0-4f2d-a1d8-e8117872b9f6","Type":"ContainerStarted","Data":"872b7cec3eb958c6eb5be593f58f6ba4944216d4ac378df1f0cb089a0a5c53f4"} Nov 24 19:19:11 crc kubenswrapper[5035]: I1124 19:19:11.024891 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tr6hr" podStartSLOduration=2.8993329 podStartE2EDuration="6.024873306s" podCreationTimestamp="2025-11-24 19:19:05 +0000 UTC" firstStartedPulling="2025-11-24 19:19:06.956253863 +0000 UTC m=+285.478760120" lastFinishedPulling="2025-11-24 19:19:10.081794269 +0000 UTC m=+288.604300526" observedRunningTime="2025-11-24 19:19:11.023112858 +0000 UTC m=+289.545619115" watchObservedRunningTime="2025-11-24 19:19:11.024873306 +0000 UTC m=+289.547379563" Nov 24 19:19:12 crc kubenswrapper[5035]: I1124 19:19:12.356529 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:12 crc kubenswrapper[5035]: I1124 19:19:12.356896 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:12 crc kubenswrapper[5035]: I1124 19:19:12.419481 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:13 crc kubenswrapper[5035]: I1124 19:19:13.068542 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 19:19:13 crc kubenswrapper[5035]: I1124 19:19:13.415124 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:13 crc kubenswrapper[5035]: I1124 19:19:13.415440 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:13 crc kubenswrapper[5035]: I1124 19:19:13.453952 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:14 crc kubenswrapper[5035]: I1124 19:19:14.055652 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x2vrj" Nov 24 19:19:14 crc kubenswrapper[5035]: I1124 19:19:14.785624 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:14 crc kubenswrapper[5035]: I1124 19:19:14.786041 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:14 crc kubenswrapper[5035]: I1124 19:19:14.826646 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:15 crc kubenswrapper[5035]: I1124 19:19:15.064773 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:19:15 crc kubenswrapper[5035]: I1124 19:19:15.785274 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:15 crc kubenswrapper[5035]: I1124 19:19:15.785330 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:15 crc kubenswrapper[5035]: I1124 19:19:15.823560 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:19:16 crc kubenswrapper[5035]: I1124 19:19:16.072788 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tr6hr" Nov 24 19:20:45 crc kubenswrapper[5035]: I1124 19:20:45.234696 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:20:45 crc kubenswrapper[5035]: I1124 19:20:45.235395 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:21:15 crc kubenswrapper[5035]: I1124 19:21:15.235130 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:21:15 crc kubenswrapper[5035]: I1124 19:21:15.235776 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.234811 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.235361 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.235407 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.235806 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71af2052370dcb5cad38af3277bc18c2dd745f2273be300ec615c3e8cdb031fa"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.235865 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://71af2052370dcb5cad38af3277bc18c2dd745f2273be300ec615c3e8cdb031fa" gracePeriod=600 Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.916208 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="71af2052370dcb5cad38af3277bc18c2dd745f2273be300ec615c3e8cdb031fa" exitCode=0 Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.916248 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"71af2052370dcb5cad38af3277bc18c2dd745f2273be300ec615c3e8cdb031fa"} Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.916786 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"b978ecf55cf22195f7615a97b366d61585baa053ee9c7eac1b8d508bebaa7f5e"} Nov 24 19:21:45 crc kubenswrapper[5035]: I1124 19:21:45.916813 5035 scope.go:117] "RemoveContainer" containerID="0d8885deb7402934bf7eac0e65bc7104b90468764aeff6f240f93558671765e3" Nov 24 19:22:22 crc kubenswrapper[5035]: I1124 19:22:22.850017 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p9jr7"] Nov 24 19:22:22 crc kubenswrapper[5035]: I1124 19:22:22.851792 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:22 crc kubenswrapper[5035]: I1124 19:22:22.864544 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p9jr7"] Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.022656 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-registry-certificates\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.022749 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-bound-sa-token\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.022825 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.022948 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-trusted-ca\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.023028 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.023146 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-registry-tls\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.023699 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.023790 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg562\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-kube-api-access-dg562\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.053603 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.125036 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-registry-certificates\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.125079 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-bound-sa-token\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.125121 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-trusted-ca\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.125143 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.125171 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-registry-tls\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.125187 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.125209 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg562\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-kube-api-access-dg562\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.126234 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-registry-certificates\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.126374 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-trusted-ca\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.126417 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.133023 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.133072 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-registry-tls\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.140691 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-bound-sa-token\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.141958 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg562\" (UniqueName: \"kubernetes.io/projected/6bdf344b-0ad5-4a3d-8749-1d5af94aa389-kube-api-access-dg562\") pod \"image-registry-66df7c8f76-p9jr7\" (UID: \"6bdf344b-0ad5-4a3d-8749-1d5af94aa389\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.176079 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:23 crc kubenswrapper[5035]: I1124 19:22:23.623956 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p9jr7"] Nov 24 19:22:24 crc kubenswrapper[5035]: I1124 19:22:24.168494 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" event={"ID":"6bdf344b-0ad5-4a3d-8749-1d5af94aa389","Type":"ContainerStarted","Data":"9323f1e58538b6110ef2d3a6c466e09d02ac361900018e11bc1b1d7a606e1625"} Nov 24 19:22:24 crc kubenswrapper[5035]: I1124 19:22:24.169602 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" event={"ID":"6bdf344b-0ad5-4a3d-8749-1d5af94aa389","Type":"ContainerStarted","Data":"833b1ae48db29778eb2ae312de930600fbf92687fd47cc8a465227477f57651a"} Nov 24 19:22:24 crc kubenswrapper[5035]: I1124 19:22:24.169690 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:24 crc kubenswrapper[5035]: I1124 19:22:24.189841 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" podStartSLOduration=2.189822486 podStartE2EDuration="2.189822486s" podCreationTimestamp="2025-11-24 19:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:22:24.188191222 +0000 UTC m=+482.710697479" watchObservedRunningTime="2025-11-24 19:22:24.189822486 +0000 UTC m=+482.712328753" Nov 24 19:22:43 crc kubenswrapper[5035]: I1124 19:22:43.185181 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-p9jr7" Nov 24 19:22:43 crc kubenswrapper[5035]: I1124 19:22:43.242759 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zv4sz"] Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.291971 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" podUID="807e010a-4744-4b1d-aa7c-9cd3360e19af" containerName="registry" containerID="cri-o://fb6f9a780b5da6f25a60fc89d5e1d327ae8b2f53f39ddccb3f30e2bac5b468f6" gracePeriod=30 Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.434329 5035 generic.go:334] "Generic (PLEG): container finished" podID="807e010a-4744-4b1d-aa7c-9cd3360e19af" containerID="fb6f9a780b5da6f25a60fc89d5e1d327ae8b2f53f39ddccb3f30e2bac5b468f6" exitCode=0 Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.434376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" event={"ID":"807e010a-4744-4b1d-aa7c-9cd3360e19af","Type":"ContainerDied","Data":"fb6f9a780b5da6f25a60fc89d5e1d327ae8b2f53f39ddccb3f30e2bac5b468f6"} Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.638124 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.768709 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-certificates\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.768771 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z96v\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-kube-api-access-6z96v\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.768795 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-trusted-ca\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.768824 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/807e010a-4744-4b1d-aa7c-9cd3360e19af-ca-trust-extracted\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.768873 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/807e010a-4744-4b1d-aa7c-9cd3360e19af-installation-pull-secrets\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.768897 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-tls\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.769048 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.769076 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-bound-sa-token\") pod \"807e010a-4744-4b1d-aa7c-9cd3360e19af\" (UID: \"807e010a-4744-4b1d-aa7c-9cd3360e19af\") " Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.769854 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.769879 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.774868 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.775300 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807e010a-4744-4b1d-aa7c-9cd3360e19af-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.775945 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-kube-api-access-6z96v" (OuterVolumeSpecName: "kube-api-access-6z96v") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "kube-api-access-6z96v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.780630 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.780639 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.785936 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/807e010a-4744-4b1d-aa7c-9cd3360e19af-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "807e010a-4744-4b1d-aa7c-9cd3360e19af" (UID: "807e010a-4744-4b1d-aa7c-9cd3360e19af"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.870235 5035 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.870283 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z96v\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-kube-api-access-6z96v\") on node \"crc\" DevicePath \"\"" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.870331 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/807e010a-4744-4b1d-aa7c-9cd3360e19af-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.870348 5035 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/807e010a-4744-4b1d-aa7c-9cd3360e19af-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.870360 5035 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/807e010a-4744-4b1d-aa7c-9cd3360e19af-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.870372 5035 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:23:08 crc kubenswrapper[5035]: I1124 19:23:08.870383 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/807e010a-4744-4b1d-aa7c-9cd3360e19af-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 19:23:09 crc kubenswrapper[5035]: I1124 19:23:09.442910 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" event={"ID":"807e010a-4744-4b1d-aa7c-9cd3360e19af","Type":"ContainerDied","Data":"891281a9f0fe5e3b6547ce4fcae0da7ccdd0f6291c97c4189db59f5e41dcefa8"} Nov 24 19:23:09 crc kubenswrapper[5035]: I1124 19:23:09.442973 5035 scope.go:117] "RemoveContainer" containerID="fb6f9a780b5da6f25a60fc89d5e1d327ae8b2f53f39ddccb3f30e2bac5b468f6" Nov 24 19:23:09 crc kubenswrapper[5035]: I1124 19:23:09.442986 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zv4sz" Nov 24 19:23:09 crc kubenswrapper[5035]: I1124 19:23:09.472805 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zv4sz"] Nov 24 19:23:09 crc kubenswrapper[5035]: I1124 19:23:09.480686 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zv4sz"] Nov 24 19:23:10 crc kubenswrapper[5035]: I1124 19:23:10.211999 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="807e010a-4744-4b1d-aa7c-9cd3360e19af" path="/var/lib/kubelet/pods/807e010a-4744-4b1d-aa7c-9cd3360e19af/volumes" Nov 24 19:23:45 crc kubenswrapper[5035]: I1124 19:23:45.234949 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:23:45 crc kubenswrapper[5035]: I1124 19:23:45.235616 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:24:15 crc kubenswrapper[5035]: I1124 19:24:15.234262 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:24:15 crc kubenswrapper[5035]: I1124 19:24:15.234855 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:24:45 crc kubenswrapper[5035]: I1124 19:24:45.234879 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:24:45 crc kubenswrapper[5035]: I1124 19:24:45.235655 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:24:45 crc kubenswrapper[5035]: I1124 19:24:45.235727 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:24:45 crc kubenswrapper[5035]: I1124 19:24:45.236670 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b978ecf55cf22195f7615a97b366d61585baa053ee9c7eac1b8d508bebaa7f5e"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:24:45 crc kubenswrapper[5035]: I1124 19:24:45.236781 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://b978ecf55cf22195f7615a97b366d61585baa053ee9c7eac1b8d508bebaa7f5e" gracePeriod=600 Nov 24 19:24:46 crc kubenswrapper[5035]: I1124 19:24:46.337965 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="b978ecf55cf22195f7615a97b366d61585baa053ee9c7eac1b8d508bebaa7f5e" exitCode=0 Nov 24 19:24:46 crc kubenswrapper[5035]: I1124 19:24:46.338163 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"b978ecf55cf22195f7615a97b366d61585baa053ee9c7eac1b8d508bebaa7f5e"} Nov 24 19:24:46 crc kubenswrapper[5035]: I1124 19:24:46.338609 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"88531e6ce05adbac2436f424c2cf6aa59384e9fc263400422a229df68ffdfb39"} Nov 24 19:24:46 crc kubenswrapper[5035]: I1124 19:24:46.338647 5035 scope.go:117] "RemoveContainer" containerID="71af2052370dcb5cad38af3277bc18c2dd745f2273be300ec615c3e8cdb031fa" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.887062 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qbfvl"] Nov 24 19:25:43 crc kubenswrapper[5035]: E1124 19:25:43.887938 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="807e010a-4744-4b1d-aa7c-9cd3360e19af" containerName="registry" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.887956 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="807e010a-4744-4b1d-aa7c-9cd3360e19af" containerName="registry" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.888084 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="807e010a-4744-4b1d-aa7c-9cd3360e19af" containerName="registry" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.888618 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.890806 5035 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-ntlcz" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.890956 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.891991 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.892265 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nxkrb"] Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.893143 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nxkrb" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.896897 5035 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-rj4s9" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.899922 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nxkrb"] Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.906719 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-kktnk"] Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.908900 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.910530 5035 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-zbw5d" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.940786 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qbfvl"] Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.952724 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-kktnk"] Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.987082 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmdvw\" (UniqueName: \"kubernetes.io/projected/d1e84c7b-08b8-411b-af66-5b1dd7f7f520-kube-api-access-mmdvw\") pod \"cert-manager-5b446d88c5-nxkrb\" (UID: \"d1e84c7b-08b8-411b-af66-5b1dd7f7f520\") " pod="cert-manager/cert-manager-5b446d88c5-nxkrb" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.987172 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8csts\" (UniqueName: \"kubernetes.io/projected/12d8a7dc-4150-452b-8c34-121cb6e65810-kube-api-access-8csts\") pod \"cert-manager-cainjector-7f985d654d-qbfvl\" (UID: \"12d8a7dc-4150-452b-8c34-121cb6e65810\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" Nov 24 19:25:43 crc kubenswrapper[5035]: I1124 19:25:43.987209 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2gth\" (UniqueName: \"kubernetes.io/projected/2dc727b8-eefd-4ce5-bae3-ab99280b162a-kube-api-access-v2gth\") pod \"cert-manager-webhook-5655c58dd6-kktnk\" (UID: \"2dc727b8-eefd-4ce5-bae3-ab99280b162a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.088050 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmdvw\" (UniqueName: \"kubernetes.io/projected/d1e84c7b-08b8-411b-af66-5b1dd7f7f520-kube-api-access-mmdvw\") pod \"cert-manager-5b446d88c5-nxkrb\" (UID: \"d1e84c7b-08b8-411b-af66-5b1dd7f7f520\") " pod="cert-manager/cert-manager-5b446d88c5-nxkrb" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.088110 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8csts\" (UniqueName: \"kubernetes.io/projected/12d8a7dc-4150-452b-8c34-121cb6e65810-kube-api-access-8csts\") pod \"cert-manager-cainjector-7f985d654d-qbfvl\" (UID: \"12d8a7dc-4150-452b-8c34-121cb6e65810\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.088143 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2gth\" (UniqueName: \"kubernetes.io/projected/2dc727b8-eefd-4ce5-bae3-ab99280b162a-kube-api-access-v2gth\") pod \"cert-manager-webhook-5655c58dd6-kktnk\" (UID: \"2dc727b8-eefd-4ce5-bae3-ab99280b162a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.108335 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2gth\" (UniqueName: \"kubernetes.io/projected/2dc727b8-eefd-4ce5-bae3-ab99280b162a-kube-api-access-v2gth\") pod \"cert-manager-webhook-5655c58dd6-kktnk\" (UID: \"2dc727b8-eefd-4ce5-bae3-ab99280b162a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.112333 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmdvw\" (UniqueName: \"kubernetes.io/projected/d1e84c7b-08b8-411b-af66-5b1dd7f7f520-kube-api-access-mmdvw\") pod \"cert-manager-5b446d88c5-nxkrb\" (UID: \"d1e84c7b-08b8-411b-af66-5b1dd7f7f520\") " pod="cert-manager/cert-manager-5b446d88c5-nxkrb" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.113860 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8csts\" (UniqueName: \"kubernetes.io/projected/12d8a7dc-4150-452b-8c34-121cb6e65810-kube-api-access-8csts\") pod \"cert-manager-cainjector-7f985d654d-qbfvl\" (UID: \"12d8a7dc-4150-452b-8c34-121cb6e65810\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.212442 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.230530 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nxkrb" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.241046 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.431924 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nxkrb"] Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.451220 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.715367 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nxkrb" event={"ID":"d1e84c7b-08b8-411b-af66-5b1dd7f7f520","Type":"ContainerStarted","Data":"4ebe733be1b7e71f52fe6c96767e2ddf03eaa4ebed8b9ba005ed5d84b1119357"} Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.751701 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-kktnk"] Nov 24 19:25:44 crc kubenswrapper[5035]: I1124 19:25:44.754206 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qbfvl"] Nov 24 19:25:44 crc kubenswrapper[5035]: W1124 19:25:44.756085 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12d8a7dc_4150_452b_8c34_121cb6e65810.slice/crio-5f0e3c04fd49aad2081913d41b9349718ec7bf49b9bf5eea6970010c65d6d912 WatchSource:0}: Error finding container 5f0e3c04fd49aad2081913d41b9349718ec7bf49b9bf5eea6970010c65d6d912: Status 404 returned error can't find the container with id 5f0e3c04fd49aad2081913d41b9349718ec7bf49b9bf5eea6970010c65d6d912 Nov 24 19:25:44 crc kubenswrapper[5035]: W1124 19:25:44.759699 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dc727b8_eefd_4ce5_bae3_ab99280b162a.slice/crio-7e3476ca1db32a27d9e4e2ba596185ceb33cc55cd5abac122db029140d4d99ee WatchSource:0}: Error finding container 7e3476ca1db32a27d9e4e2ba596185ceb33cc55cd5abac122db029140d4d99ee: Status 404 returned error can't find the container with id 7e3476ca1db32a27d9e4e2ba596185ceb33cc55cd5abac122db029140d4d99ee Nov 24 19:25:45 crc kubenswrapper[5035]: I1124 19:25:45.722864 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" event={"ID":"12d8a7dc-4150-452b-8c34-121cb6e65810","Type":"ContainerStarted","Data":"5f0e3c04fd49aad2081913d41b9349718ec7bf49b9bf5eea6970010c65d6d912"} Nov 24 19:25:45 crc kubenswrapper[5035]: I1124 19:25:45.723809 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" event={"ID":"2dc727b8-eefd-4ce5-bae3-ab99280b162a","Type":"ContainerStarted","Data":"7e3476ca1db32a27d9e4e2ba596185ceb33cc55cd5abac122db029140d4d99ee"} Nov 24 19:25:48 crc kubenswrapper[5035]: I1124 19:25:48.744262 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nxkrb" event={"ID":"d1e84c7b-08b8-411b-af66-5b1dd7f7f520","Type":"ContainerStarted","Data":"4c4ff79213a02a0698b7bbeecf93a589f84b8b3a56238670d988ec0484e894f6"} Nov 24 19:25:48 crc kubenswrapper[5035]: I1124 19:25:48.760849 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-nxkrb" podStartSLOduration=1.653840854 podStartE2EDuration="5.760829084s" podCreationTimestamp="2025-11-24 19:25:43 +0000 UTC" firstStartedPulling="2025-11-24 19:25:44.451047211 +0000 UTC m=+682.973553468" lastFinishedPulling="2025-11-24 19:25:48.558035431 +0000 UTC m=+687.080541698" observedRunningTime="2025-11-24 19:25:48.759875848 +0000 UTC m=+687.282382115" watchObservedRunningTime="2025-11-24 19:25:48.760829084 +0000 UTC m=+687.283335361" Nov 24 19:25:50 crc kubenswrapper[5035]: I1124 19:25:50.759204 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" event={"ID":"12d8a7dc-4150-452b-8c34-121cb6e65810","Type":"ContainerStarted","Data":"793a18c777d30c58407f555c350dd7319ca21edd2e7e9bd1dc16d1662851fe09"} Nov 24 19:25:50 crc kubenswrapper[5035]: I1124 19:25:50.762821 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" event={"ID":"2dc727b8-eefd-4ce5-bae3-ab99280b162a","Type":"ContainerStarted","Data":"afaf8a1a4cc92ccf0e4db1845304d38da320c14ebde8b349d52e7de61543deb0"} Nov 24 19:25:50 crc kubenswrapper[5035]: I1124 19:25:50.762997 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" Nov 24 19:25:50 crc kubenswrapper[5035]: I1124 19:25:50.780032 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qbfvl" podStartSLOduration=2.708087735 podStartE2EDuration="7.779865033s" podCreationTimestamp="2025-11-24 19:25:43 +0000 UTC" firstStartedPulling="2025-11-24 19:25:44.758951478 +0000 UTC m=+683.281457745" lastFinishedPulling="2025-11-24 19:25:49.830728786 +0000 UTC m=+688.353235043" observedRunningTime="2025-11-24 19:25:50.778128146 +0000 UTC m=+689.300634443" watchObservedRunningTime="2025-11-24 19:25:50.779865033 +0000 UTC m=+689.302371290" Nov 24 19:25:50 crc kubenswrapper[5035]: I1124 19:25:50.798084 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" podStartSLOduration=2.515412279 podStartE2EDuration="7.798049812s" podCreationTimestamp="2025-11-24 19:25:43 +0000 UTC" firstStartedPulling="2025-11-24 19:25:44.76522824 +0000 UTC m=+683.287734497" lastFinishedPulling="2025-11-24 19:25:50.047865773 +0000 UTC m=+688.570372030" observedRunningTime="2025-11-24 19:25:50.790864995 +0000 UTC m=+689.313371262" watchObservedRunningTime="2025-11-24 19:25:50.798049812 +0000 UTC m=+689.320556109" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.326087 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dbbzx"] Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.327207 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-controller" containerID="cri-o://c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.327268 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="nbdb" containerID="cri-o://838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.327402 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.327450 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-acl-logging" containerID="cri-o://3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.327433 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-node" containerID="cri-o://e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.327680 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="northd" containerID="cri-o://e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.327743 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="sbdb" containerID="cri-o://cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.371085 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" containerID="cri-o://e8ead8c9c31b17a0cb024c98dd49f1f9c8366067fc37af3c460a0a2e98a35619" gracePeriod=30 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.791246 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovnkube-controller/3.log" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.794546 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovn-acl-logging/0.log" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.795401 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovn-controller/0.log" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796513 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="e8ead8c9c31b17a0cb024c98dd49f1f9c8366067fc37af3c460a0a2e98a35619" exitCode=0 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796559 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72" exitCode=0 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796576 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721" exitCode=0 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796591 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f" exitCode=0 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796604 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746" exitCode=0 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796616 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8" exitCode=0 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796628 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045" exitCode=143 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796625 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"e8ead8c9c31b17a0cb024c98dd49f1f9c8366067fc37af3c460a0a2e98a35619"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796706 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796786 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796804 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796819 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796837 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796854 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796870 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796759 5035 scope.go:117] "RemoveContainer" containerID="6ed15f2a4f541bc2fcd52bd0dbcac2d1986fb5beddff36d136b044e0428a095d" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.796641 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerID="c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1" exitCode=143 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.800455 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/2.log" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.801169 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/1.log" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.801200 5035 generic.go:334] "Generic (PLEG): container finished" podID="38b324c3-dc7d-4555-ac0d-714a9f1a40cf" containerID="01506ee8a33f0daf99578100ec1724d23ab8e5d245387a0430b41fa3e72a0a19" exitCode=2 Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.801223 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerDied","Data":"01506ee8a33f0daf99578100ec1724d23ab8e5d245387a0430b41fa3e72a0a19"} Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.801636 5035 scope.go:117] "RemoveContainer" containerID="01506ee8a33f0daf99578100ec1724d23ab8e5d245387a0430b41fa3e72a0a19" Nov 24 19:25:54 crc kubenswrapper[5035]: E1124 19:25:54.801810 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mg7qv_openshift-multus(38b324c3-dc7d-4555-ac0d-714a9f1a40cf)\"" pod="openshift-multus/multus-mg7qv" podUID="38b324c3-dc7d-4555-ac0d-714a9f1a40cf" Nov 24 19:25:54 crc kubenswrapper[5035]: I1124 19:25:54.863283 5035 scope.go:117] "RemoveContainer" containerID="91c651986e7ada079370ae42bf71fe6894655510fa0b979cbecf5ef057c89965" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.097596 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovn-acl-logging/0.log" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.098522 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovn-controller/0.log" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.099045 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158125 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wqk95"] Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158454 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-acl-logging" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158487 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-acl-logging" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158508 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158520 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158538 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158552 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158578 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kubecfg-setup" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158590 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kubecfg-setup" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158606 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158617 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158636 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="nbdb" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158652 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="nbdb" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158687 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-node" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158703 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-node" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158719 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="sbdb" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158738 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="sbdb" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158759 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158772 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158790 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158802 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158820 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="northd" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158832 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="northd" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.158844 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.158856 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159018 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159035 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159055 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-acl-logging" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159071 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-node" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159088 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="northd" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159102 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="nbdb" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159114 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="sbdb" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159127 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159142 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159156 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovn-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159172 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 19:25:55 crc kubenswrapper[5035]: E1124 19:25:55.159355 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159369 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.159594 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" containerName="ovnkube-controller" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.162576 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.259898 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-etc-openvswitch\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.259985 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-config\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260038 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-slash\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260032 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260074 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-systemd\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260102 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-bin\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260116 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-slash" (OuterVolumeSpecName: "host-slash") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260141 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-env-overrides\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260159 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260173 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-script-lib\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260204 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-kubelet\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260233 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-ovn-kubernetes\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260265 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbkht\" (UniqueName: \"kubernetes.io/projected/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-kube-api-access-lbkht\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260323 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-openvswitch\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260364 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-systemd-units\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260321 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260340 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260394 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260402 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260474 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260499 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-node-log\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260521 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260527 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-netd\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260552 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-node-log" (OuterVolumeSpecName: "node-log") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260561 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-ovn\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260579 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260592 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovn-node-metrics-cert\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260605 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260615 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-var-lib-openvswitch\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260633 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-log-socket\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260647 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-netns\") pod \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\" (UID: \"4d9cb31b-5f2a-4594-ad12-718b6e99d15f\") " Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260652 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260688 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260710 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-log-socket" (OuterVolumeSpecName: "log-socket") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260729 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260801 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.260873 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261200 5035 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261231 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261255 5035 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261276 5035 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261345 5035 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261376 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261399 5035 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261422 5035 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261447 5035 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261471 5035 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261497 5035 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261521 5035 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261545 5035 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261571 5035 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261592 5035 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261614 5035 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.261636 5035 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.265684 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.266179 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-kube-api-access-lbkht" (OuterVolumeSpecName: "kube-api-access-lbkht") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "kube-api-access-lbkht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.272899 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "4d9cb31b-5f2a-4594-ad12-718b6e99d15f" (UID: "4d9cb31b-5f2a-4594-ad12-718b6e99d15f"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362789 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-systemd\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362839 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-env-overrides\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362863 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-ovn\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362879 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-slash\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362893 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362911 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-systemd-units\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362928 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovnkube-config\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362947 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-kubelet\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362961 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovn-node-metrics-cert\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.362982 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363011 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-run-netns\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363039 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-node-log\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363098 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xqnb\" (UniqueName: \"kubernetes.io/projected/a0fcf5ed-9055-402a-86c5-e0e4256e4162-kube-api-access-6xqnb\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363281 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovnkube-script-lib\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363365 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-log-socket\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363383 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-etc-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363398 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-var-lib-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363439 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-cni-bin\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363500 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-cni-netd\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363530 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-run-ovn-kubernetes\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363568 5035 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363580 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbkht\" (UniqueName: \"kubernetes.io/projected/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-kube-api-access-lbkht\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.363589 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4d9cb31b-5f2a-4594-ad12-718b6e99d15f-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464141 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-cni-netd\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464189 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-run-ovn-kubernetes\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464216 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-systemd\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464235 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-env-overrides\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464253 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-ovn\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464270 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464285 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-slash\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464336 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-systemd\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464362 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-systemd-units\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464340 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-systemd-units\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464394 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464376 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-run-ovn\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464432 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-run-ovn-kubernetes\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464415 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-slash\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464404 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovnkube-config\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464467 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-kubelet\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464484 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovn-node-metrics-cert\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464500 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464519 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-node-log\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464537 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-run-netns\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464561 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xqnb\" (UniqueName: \"kubernetes.io/projected/a0fcf5ed-9055-402a-86c5-e0e4256e4162-kube-api-access-6xqnb\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464587 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovnkube-script-lib\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464594 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464615 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-log-socket\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464637 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-kubelet\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464641 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-etc-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464667 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-etc-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464677 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-var-lib-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464698 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-node-log\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464704 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-cni-bin\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464729 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-run-netns\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464761 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-cni-bin\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.465123 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-env-overrides\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.465180 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-var-lib-openvswitch\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.464338 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-host-cni-netd\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.465230 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a0fcf5ed-9055-402a-86c5-e0e4256e4162-log-socket\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.465384 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovnkube-config\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.465544 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovnkube-script-lib\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.467583 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a0fcf5ed-9055-402a-86c5-e0e4256e4162-ovn-node-metrics-cert\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.494204 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xqnb\" (UniqueName: \"kubernetes.io/projected/a0fcf5ed-9055-402a-86c5-e0e4256e4162-kube-api-access-6xqnb\") pod \"ovnkube-node-wqk95\" (UID: \"a0fcf5ed-9055-402a-86c5-e0e4256e4162\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.786949 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.819680 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovn-acl-logging/0.log" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.822273 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dbbzx_4d9cb31b-5f2a-4594-ad12-718b6e99d15f/ovn-controller/0.log" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.825550 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" event={"ID":"4d9cb31b-5f2a-4594-ad12-718b6e99d15f","Type":"ContainerDied","Data":"2c580bb5b6b4c531b89e2ac918b89057fd8714911f549d6138246fbbddd1d910"} Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.825611 5035 scope.go:117] "RemoveContainer" containerID="e8ead8c9c31b17a0cb024c98dd49f1f9c8366067fc37af3c460a0a2e98a35619" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.825784 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dbbzx" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.837030 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/2.log" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.861791 5035 scope.go:117] "RemoveContainer" containerID="cee99ada612d7a83134e4417f8f5a22c8e340b8abbedf57f97e4694743265b72" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.892071 5035 scope.go:117] "RemoveContainer" containerID="838339115c2b5fc746d3e9f80aca5846192ad7c9c52d692f8f1bc88fbe34c721" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.896335 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dbbzx"] Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.907652 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dbbzx"] Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.916146 5035 scope.go:117] "RemoveContainer" containerID="e76b62907da9d3fb0e2b5eeb4b5aadaaff9e2dd0cdb618571f5523b8c5ce3c4f" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.927725 5035 scope.go:117] "RemoveContainer" containerID="f3ee5b66167a8bfaadf99fcb9b9decdebd3b8b220c8f00167ea79bda71413746" Nov 24 19:25:55 crc kubenswrapper[5035]: I1124 19:25:55.943558 5035 scope.go:117] "RemoveContainer" containerID="e2c985dcdccb8e9d4f2006a1c09d876f096b1936916abd8e2236d3ca1c57d5b8" Nov 24 19:25:56 crc kubenswrapper[5035]: I1124 19:25:56.013430 5035 scope.go:117] "RemoveContainer" containerID="3fa2f6c94f93f3c8adfab548b95b916277c54bfc23ee5eb281090784039c6045" Nov 24 19:25:56 crc kubenswrapper[5035]: I1124 19:25:56.033213 5035 scope.go:117] "RemoveContainer" containerID="c52e81555170e7939be22f54c79eaff08c667bdd926d6f9a57bf61138a07f9d1" Nov 24 19:25:56 crc kubenswrapper[5035]: I1124 19:25:56.048509 5035 scope.go:117] "RemoveContainer" containerID="baa8e3aa8d63ccfe2367f613e869d424325a426cb935bde7a8a6b69f215d1ba3" Nov 24 19:25:56 crc kubenswrapper[5035]: I1124 19:25:56.214851 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d9cb31b-5f2a-4594-ad12-718b6e99d15f" path="/var/lib/kubelet/pods/4d9cb31b-5f2a-4594-ad12-718b6e99d15f/volumes" Nov 24 19:25:56 crc kubenswrapper[5035]: I1124 19:25:56.846476 5035 generic.go:334] "Generic (PLEG): container finished" podID="a0fcf5ed-9055-402a-86c5-e0e4256e4162" containerID="b95aa9f6f4ed749ed0d9225bf593c5fe6e4cf469501e9e6a2b1ab5e6d815d049" exitCode=0 Nov 24 19:25:56 crc kubenswrapper[5035]: I1124 19:25:56.846556 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerDied","Data":"b95aa9f6f4ed749ed0d9225bf593c5fe6e4cf469501e9e6a2b1ab5e6d815d049"} Nov 24 19:25:56 crc kubenswrapper[5035]: I1124 19:25:56.846618 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"596951abea973b0a86e2fd494d1ef8f7a998a7b0be3d899b01d597ded9a1818f"} Nov 24 19:25:57 crc kubenswrapper[5035]: I1124 19:25:57.868134 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"4e28493f09cd14bc98784da50f8d272730b312842b68349545a828e898cc90d9"} Nov 24 19:25:57 crc kubenswrapper[5035]: I1124 19:25:57.868716 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"ddf12cde21e9624f53249ebe24aefa64c8575fafcbbfc9025ee391990f640b22"} Nov 24 19:25:57 crc kubenswrapper[5035]: I1124 19:25:57.868732 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"430777019b8c9bb6f217219c8961bb5137cba66f87b1d1d709b3b42cf1e3dd50"} Nov 24 19:25:57 crc kubenswrapper[5035]: I1124 19:25:57.868745 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"76db6a40f8fee1253812f68709178aa821f2def83c603c94d91172bc1bfa95b8"} Nov 24 19:25:57 crc kubenswrapper[5035]: I1124 19:25:57.868755 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"0e160a129876b8c0b768be4b4ff3ae0298c4a80e494e60829d485e565693a9c7"} Nov 24 19:25:57 crc kubenswrapper[5035]: I1124 19:25:57.868764 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"f5b37124f4707130f9015db488402f9ad2d4524c333ee47ca91464490ad808a2"} Nov 24 19:25:59 crc kubenswrapper[5035]: I1124 19:25:59.245351 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-kktnk" Nov 24 19:26:00 crc kubenswrapper[5035]: I1124 19:26:00.889846 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"f0107b5dd2126a913dacae3b7b291ed409301144a8d88c495a3eadc97746ccda"} Nov 24 19:26:02 crc kubenswrapper[5035]: I1124 19:26:02.903747 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" event={"ID":"a0fcf5ed-9055-402a-86c5-e0e4256e4162","Type":"ContainerStarted","Data":"d5bb05b0e5945d278b6a270f8a55d2567f21d4a739f43569995fe40bb3d8c8ad"} Nov 24 19:26:02 crc kubenswrapper[5035]: I1124 19:26:02.904229 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:26:02 crc kubenswrapper[5035]: I1124 19:26:02.904247 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:26:02 crc kubenswrapper[5035]: I1124 19:26:02.904257 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:26:02 crc kubenswrapper[5035]: I1124 19:26:02.935111 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" podStartSLOduration=7.935090745 podStartE2EDuration="7.935090745s" podCreationTimestamp="2025-11-24 19:25:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:26:02.931486126 +0000 UTC m=+701.453992393" watchObservedRunningTime="2025-11-24 19:26:02.935090745 +0000 UTC m=+701.457597002" Nov 24 19:26:02 crc kubenswrapper[5035]: I1124 19:26:02.935964 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:26:02 crc kubenswrapper[5035]: I1124 19:26:02.943676 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:26:05 crc kubenswrapper[5035]: I1124 19:26:05.200758 5035 scope.go:117] "RemoveContainer" containerID="01506ee8a33f0daf99578100ec1724d23ab8e5d245387a0430b41fa3e72a0a19" Nov 24 19:26:05 crc kubenswrapper[5035]: E1124 19:26:05.201536 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mg7qv_openshift-multus(38b324c3-dc7d-4555-ac0d-714a9f1a40cf)\"" pod="openshift-multus/multus-mg7qv" podUID="38b324c3-dc7d-4555-ac0d-714a9f1a40cf" Nov 24 19:26:18 crc kubenswrapper[5035]: I1124 19:26:18.200419 5035 scope.go:117] "RemoveContainer" containerID="01506ee8a33f0daf99578100ec1724d23ab8e5d245387a0430b41fa3e72a0a19" Nov 24 19:26:19 crc kubenswrapper[5035]: I1124 19:26:19.008992 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mg7qv_38b324c3-dc7d-4555-ac0d-714a9f1a40cf/kube-multus/2.log" Nov 24 19:26:19 crc kubenswrapper[5035]: I1124 19:26:19.009409 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mg7qv" event={"ID":"38b324c3-dc7d-4555-ac0d-714a9f1a40cf","Type":"ContainerStarted","Data":"b030d56686edaaa3d6688c15fb8233310fa4c82e93c638979a22aac76f5d0c60"} Nov 24 19:26:25 crc kubenswrapper[5035]: I1124 19:26:25.822236 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wqk95" Nov 24 19:26:45 crc kubenswrapper[5035]: I1124 19:26:45.234809 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:26:45 crc kubenswrapper[5035]: I1124 19:26:45.235338 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.233166 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr"] Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.235135 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr"] Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.235258 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.240129 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.345889 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxqxc\" (UniqueName: \"kubernetes.io/projected/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-kube-api-access-nxqxc\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.346081 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.346779 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.448007 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.448246 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.448337 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxqxc\" (UniqueName: \"kubernetes.io/projected/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-kube-api-access-nxqxc\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.448520 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.449894 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.473459 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxqxc\" (UniqueName: \"kubernetes.io/projected/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-kube-api-access-nxqxc\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.553030 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:48 crc kubenswrapper[5035]: I1124 19:26:48.958605 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr"] Nov 24 19:26:49 crc kubenswrapper[5035]: I1124 19:26:49.179363 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" event={"ID":"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc","Type":"ContainerStarted","Data":"12dc3800a55d75634de2ef4e8c37b4fac8fa2bb20098c5cda641dde308ebf2f6"} Nov 24 19:26:49 crc kubenswrapper[5035]: I1124 19:26:49.179697 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" event={"ID":"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc","Type":"ContainerStarted","Data":"f2fa5e91c97e8805f6fa8b551d6806556eec1d16fe6b17c40ada86c9dac23b9e"} Nov 24 19:26:50 crc kubenswrapper[5035]: I1124 19:26:50.185591 5035 generic.go:334] "Generic (PLEG): container finished" podID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerID="12dc3800a55d75634de2ef4e8c37b4fac8fa2bb20098c5cda641dde308ebf2f6" exitCode=0 Nov 24 19:26:50 crc kubenswrapper[5035]: I1124 19:26:50.185671 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" event={"ID":"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc","Type":"ContainerDied","Data":"12dc3800a55d75634de2ef4e8c37b4fac8fa2bb20098c5cda641dde308ebf2f6"} Nov 24 19:26:52 crc kubenswrapper[5035]: I1124 19:26:52.198945 5035 generic.go:334] "Generic (PLEG): container finished" podID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerID="006d668ed294e69f825cf901b8a0d0cc3bcc484903cb2cc34c5df3bbda88e0fc" exitCode=0 Nov 24 19:26:52 crc kubenswrapper[5035]: I1124 19:26:52.198947 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" event={"ID":"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc","Type":"ContainerDied","Data":"006d668ed294e69f825cf901b8a0d0cc3bcc484903cb2cc34c5df3bbda88e0fc"} Nov 24 19:26:53 crc kubenswrapper[5035]: I1124 19:26:53.211857 5035 generic.go:334] "Generic (PLEG): container finished" podID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerID="4b8d945fdc57333832ae6f66f03cd47a15c7dba75b989f92defb6ed4172c394b" exitCode=0 Nov 24 19:26:53 crc kubenswrapper[5035]: I1124 19:26:53.211994 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" event={"ID":"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc","Type":"ContainerDied","Data":"4b8d945fdc57333832ae6f66f03cd47a15c7dba75b989f92defb6ed4172c394b"} Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.496384 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.631710 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-bundle\") pod \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.631948 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxqxc\" (UniqueName: \"kubernetes.io/projected/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-kube-api-access-nxqxc\") pod \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.632047 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-util\") pod \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\" (UID: \"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc\") " Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.633171 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-bundle" (OuterVolumeSpecName: "bundle") pod "efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" (UID: "efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.641811 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-kube-api-access-nxqxc" (OuterVolumeSpecName: "kube-api-access-nxqxc") pod "efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" (UID: "efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc"). InnerVolumeSpecName "kube-api-access-nxqxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.734534 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxqxc\" (UniqueName: \"kubernetes.io/projected/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-kube-api-access-nxqxc\") on node \"crc\" DevicePath \"\"" Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.734629 5035 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.813404 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-util" (OuterVolumeSpecName: "util") pod "efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" (UID: "efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:26:54 crc kubenswrapper[5035]: I1124 19:26:54.835401 5035 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc-util\") on node \"crc\" DevicePath \"\"" Nov 24 19:26:55 crc kubenswrapper[5035]: I1124 19:26:55.226766 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" event={"ID":"efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc","Type":"ContainerDied","Data":"f2fa5e91c97e8805f6fa8b551d6806556eec1d16fe6b17c40ada86c9dac23b9e"} Nov 24 19:26:55 crc kubenswrapper[5035]: I1124 19:26:55.226823 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2fa5e91c97e8805f6fa8b551d6806556eec1d16fe6b17c40ada86c9dac23b9e" Nov 24 19:26:55 crc kubenswrapper[5035]: I1124 19:26:55.226910 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.982805 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-cdj9v"] Nov 24 19:26:56 crc kubenswrapper[5035]: E1124 19:26:56.983271 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerName="extract" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.983283 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerName="extract" Nov 24 19:26:56 crc kubenswrapper[5035]: E1124 19:26:56.983312 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerName="pull" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.983317 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerName="pull" Nov 24 19:26:56 crc kubenswrapper[5035]: E1124 19:26:56.983328 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerName="util" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.983334 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerName="util" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.983422 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc" containerName="extract" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.983809 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.988713 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.988896 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 19:26:56 crc kubenswrapper[5035]: I1124 19:26:56.988999 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-wq9x7" Nov 24 19:26:57 crc kubenswrapper[5035]: I1124 19:26:57.005141 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-cdj9v"] Nov 24 19:26:57 crc kubenswrapper[5035]: I1124 19:26:57.068955 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4862q\" (UniqueName: \"kubernetes.io/projected/b0ce9ef4-de1c-41fa-b76f-2a31fc591d54-kube-api-access-4862q\") pod \"nmstate-operator-557fdffb88-cdj9v\" (UID: \"b0ce9ef4-de1c-41fa-b76f-2a31fc591d54\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" Nov 24 19:26:57 crc kubenswrapper[5035]: I1124 19:26:57.170277 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4862q\" (UniqueName: \"kubernetes.io/projected/b0ce9ef4-de1c-41fa-b76f-2a31fc591d54-kube-api-access-4862q\") pod \"nmstate-operator-557fdffb88-cdj9v\" (UID: \"b0ce9ef4-de1c-41fa-b76f-2a31fc591d54\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" Nov 24 19:26:57 crc kubenswrapper[5035]: I1124 19:26:57.190701 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4862q\" (UniqueName: \"kubernetes.io/projected/b0ce9ef4-de1c-41fa-b76f-2a31fc591d54-kube-api-access-4862q\") pod \"nmstate-operator-557fdffb88-cdj9v\" (UID: \"b0ce9ef4-de1c-41fa-b76f-2a31fc591d54\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" Nov 24 19:26:57 crc kubenswrapper[5035]: I1124 19:26:57.298206 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" Nov 24 19:26:57 crc kubenswrapper[5035]: I1124 19:26:57.506122 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-cdj9v"] Nov 24 19:26:58 crc kubenswrapper[5035]: I1124 19:26:58.243538 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" event={"ID":"b0ce9ef4-de1c-41fa-b76f-2a31fc591d54","Type":"ContainerStarted","Data":"a173ece7ad04dc06a130a50c23509510c59d25dfe7e8698b8f0dd9965bdd030d"} Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.021404 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r2vkf"] Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.021863 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" podUID="727a0b65-6c8b-477d-8743-e3bd88a7c904" containerName="controller-manager" containerID="cri-o://c801e3af33bd17691ce109e5f20a5e66e798b3ea03fa9d9e08c8abdc1fb0c638" gracePeriod=30 Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.105423 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n"] Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.105676 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" podUID="d50060f5-1b00-483a-914b-c34277882c04" containerName="route-controller-manager" containerID="cri-o://cf65f0579a5e7c5698cf54ba5f7a5f6ed89d1f87a826463853549c1d2a841d4e" gracePeriod=30 Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.257069 5035 generic.go:334] "Generic (PLEG): container finished" podID="d50060f5-1b00-483a-914b-c34277882c04" containerID="cf65f0579a5e7c5698cf54ba5f7a5f6ed89d1f87a826463853549c1d2a841d4e" exitCode=0 Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.257154 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" event={"ID":"d50060f5-1b00-483a-914b-c34277882c04","Type":"ContainerDied","Data":"cf65f0579a5e7c5698cf54ba5f7a5f6ed89d1f87a826463853549c1d2a841d4e"} Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.259034 5035 generic.go:334] "Generic (PLEG): container finished" podID="727a0b65-6c8b-477d-8743-e3bd88a7c904" containerID="c801e3af33bd17691ce109e5f20a5e66e798b3ea03fa9d9e08c8abdc1fb0c638" exitCode=0 Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.259065 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" event={"ID":"727a0b65-6c8b-477d-8743-e3bd88a7c904","Type":"ContainerDied","Data":"c801e3af33bd17691ce109e5f20a5e66e798b3ea03fa9d9e08c8abdc1fb0c638"} Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.467219 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.470865 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.528940 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85"] Nov 24 19:27:00 crc kubenswrapper[5035]: E1124 19:27:00.529217 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d50060f5-1b00-483a-914b-c34277882c04" containerName="route-controller-manager" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.529237 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d50060f5-1b00-483a-914b-c34277882c04" containerName="route-controller-manager" Nov 24 19:27:00 crc kubenswrapper[5035]: E1124 19:27:00.529257 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727a0b65-6c8b-477d-8743-e3bd88a7c904" containerName="controller-manager" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.529264 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="727a0b65-6c8b-477d-8743-e3bd88a7c904" containerName="controller-manager" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.529390 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="727a0b65-6c8b-477d-8743-e3bd88a7c904" containerName="controller-manager" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.529415 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d50060f5-1b00-483a-914b-c34277882c04" containerName="route-controller-manager" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.529809 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.536138 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85"] Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618224 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-client-ca\") pod \"d50060f5-1b00-483a-914b-c34277882c04\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618268 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-868kx\" (UniqueName: \"kubernetes.io/projected/d50060f5-1b00-483a-914b-c34277882c04-kube-api-access-868kx\") pod \"d50060f5-1b00-483a-914b-c34277882c04\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618321 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-client-ca\") pod \"727a0b65-6c8b-477d-8743-e3bd88a7c904\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618349 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-config\") pod \"d50060f5-1b00-483a-914b-c34277882c04\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618382 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727a0b65-6c8b-477d-8743-e3bd88a7c904-serving-cert\") pod \"727a0b65-6c8b-477d-8743-e3bd88a7c904\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618440 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-proxy-ca-bundles\") pod \"727a0b65-6c8b-477d-8743-e3bd88a7c904\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618487 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-config\") pod \"727a0b65-6c8b-477d-8743-e3bd88a7c904\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618514 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc4c4\" (UniqueName: \"kubernetes.io/projected/727a0b65-6c8b-477d-8743-e3bd88a7c904-kube-api-access-kc4c4\") pod \"727a0b65-6c8b-477d-8743-e3bd88a7c904\" (UID: \"727a0b65-6c8b-477d-8743-e3bd88a7c904\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618546 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50060f5-1b00-483a-914b-c34277882c04-serving-cert\") pod \"d50060f5-1b00-483a-914b-c34277882c04\" (UID: \"d50060f5-1b00-483a-914b-c34277882c04\") " Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618762 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9k7q\" (UniqueName: \"kubernetes.io/projected/c5c314e3-69e3-4810-93e1-0a945a94f959-kube-api-access-g9k7q\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618800 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5c314e3-69e3-4810-93e1-0a945a94f959-config\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618832 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5c314e3-69e3-4810-93e1-0a945a94f959-serving-cert\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.618901 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5c314e3-69e3-4810-93e1-0a945a94f959-client-ca\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.619027 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-client-ca" (OuterVolumeSpecName: "client-ca") pod "d50060f5-1b00-483a-914b-c34277882c04" (UID: "d50060f5-1b00-483a-914b-c34277882c04"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.619058 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-config" (OuterVolumeSpecName: "config") pod "d50060f5-1b00-483a-914b-c34277882c04" (UID: "d50060f5-1b00-483a-914b-c34277882c04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.619653 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "727a0b65-6c8b-477d-8743-e3bd88a7c904" (UID: "727a0b65-6c8b-477d-8743-e3bd88a7c904"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.619761 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-config" (OuterVolumeSpecName: "config") pod "727a0b65-6c8b-477d-8743-e3bd88a7c904" (UID: "727a0b65-6c8b-477d-8743-e3bd88a7c904"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.620315 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-client-ca" (OuterVolumeSpecName: "client-ca") pod "727a0b65-6c8b-477d-8743-e3bd88a7c904" (UID: "727a0b65-6c8b-477d-8743-e3bd88a7c904"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.623750 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/727a0b65-6c8b-477d-8743-e3bd88a7c904-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "727a0b65-6c8b-477d-8743-e3bd88a7c904" (UID: "727a0b65-6c8b-477d-8743-e3bd88a7c904"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.624535 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/727a0b65-6c8b-477d-8743-e3bd88a7c904-kube-api-access-kc4c4" (OuterVolumeSpecName: "kube-api-access-kc4c4") pod "727a0b65-6c8b-477d-8743-e3bd88a7c904" (UID: "727a0b65-6c8b-477d-8743-e3bd88a7c904"). InnerVolumeSpecName "kube-api-access-kc4c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.624563 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d50060f5-1b00-483a-914b-c34277882c04-kube-api-access-868kx" (OuterVolumeSpecName: "kube-api-access-868kx") pod "d50060f5-1b00-483a-914b-c34277882c04" (UID: "d50060f5-1b00-483a-914b-c34277882c04"). InnerVolumeSpecName "kube-api-access-868kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.624623 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50060f5-1b00-483a-914b-c34277882c04-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d50060f5-1b00-483a-914b-c34277882c04" (UID: "d50060f5-1b00-483a-914b-c34277882c04"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.720032 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5c314e3-69e3-4810-93e1-0a945a94f959-client-ca\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.720623 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9k7q\" (UniqueName: \"kubernetes.io/projected/c5c314e3-69e3-4810-93e1-0a945a94f959-kube-api-access-g9k7q\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.720889 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5c314e3-69e3-4810-93e1-0a945a94f959-config\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.721138 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5c314e3-69e3-4810-93e1-0a945a94f959-serving-cert\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.721579 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.721787 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-868kx\" (UniqueName: \"kubernetes.io/projected/d50060f5-1b00-483a-914b-c34277882c04-kube-api-access-868kx\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.721952 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.722093 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5c314e3-69e3-4810-93e1-0a945a94f959-config\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.722114 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c5c314e3-69e3-4810-93e1-0a945a94f959-client-ca\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.722114 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d50060f5-1b00-483a-914b-c34277882c04-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.722632 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/727a0b65-6c8b-477d-8743-e3bd88a7c904-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.722791 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.722939 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727a0b65-6c8b-477d-8743-e3bd88a7c904-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.723074 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc4c4\" (UniqueName: \"kubernetes.io/projected/727a0b65-6c8b-477d-8743-e3bd88a7c904-kube-api-access-kc4c4\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.723186 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d50060f5-1b00-483a-914b-c34277882c04-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.725811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5c314e3-69e3-4810-93e1-0a945a94f959-serving-cert\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.738224 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9k7q\" (UniqueName: \"kubernetes.io/projected/c5c314e3-69e3-4810-93e1-0a945a94f959-kube-api-access-g9k7q\") pod \"route-controller-manager-7bb98bbd74-bml85\" (UID: \"c5c314e3-69e3-4810-93e1-0a945a94f959\") " pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:00 crc kubenswrapper[5035]: I1124 19:27:00.854308 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.049908 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85"] Nov 24 19:27:01 crc kubenswrapper[5035]: W1124 19:27:01.055342 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5c314e3_69e3_4810_93e1_0a945a94f959.slice/crio-28f768036736abc3488a77d39576c67d1d2f2789312d9a5162bcf2f98a1fb5b0 WatchSource:0}: Error finding container 28f768036736abc3488a77d39576c67d1d2f2789312d9a5162bcf2f98a1fb5b0: Status 404 returned error can't find the container with id 28f768036736abc3488a77d39576c67d1d2f2789312d9a5162bcf2f98a1fb5b0 Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.272040 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" event={"ID":"d50060f5-1b00-483a-914b-c34277882c04","Type":"ContainerDied","Data":"759e37b9e45bcc763b50d722427ac4522c82991da3c178214e3e70036545044f"} Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.272083 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.272155 5035 scope.go:117] "RemoveContainer" containerID="cf65f0579a5e7c5698cf54ba5f7a5f6ed89d1f87a826463853549c1d2a841d4e" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.274764 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" event={"ID":"727a0b65-6c8b-477d-8743-e3bd88a7c904","Type":"ContainerDied","Data":"477ba616681c6cefbcae9b560c9fe2fb1a3dbc6e05a341dcfdbb44a1b4d11101"} Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.276629 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" event={"ID":"c5c314e3-69e3-4810-93e1-0a945a94f959","Type":"ContainerStarted","Data":"e5d0c22b3f4c6d97a127ef2896fda7effd0111389f45d09a1cd165824ac2887f"} Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.276662 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" event={"ID":"c5c314e3-69e3-4810-93e1-0a945a94f959","Type":"ContainerStarted","Data":"28f768036736abc3488a77d39576c67d1d2f2789312d9a5162bcf2f98a1fb5b0"} Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.276848 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.276916 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-r2vkf" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.280376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" event={"ID":"b0ce9ef4-de1c-41fa-b76f-2a31fc591d54","Type":"ContainerStarted","Data":"8f8f06b596a747efbf234cf584f4c2d68da663424b5b528807a78e9f3e9ccaf0"} Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.287589 5035 scope.go:117] "RemoveContainer" containerID="c801e3af33bd17691ce109e5f20a5e66e798b3ea03fa9d9e08c8abdc1fb0c638" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.302741 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" podStartSLOduration=1.302725741 podStartE2EDuration="1.302725741s" podCreationTimestamp="2025-11-24 19:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:27:01.299171213 +0000 UTC m=+759.821677470" watchObservedRunningTime="2025-11-24 19:27:01.302725741 +0000 UTC m=+759.825231998" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.320034 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r2vkf"] Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.322940 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-r2vkf"] Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.330879 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n"] Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.338390 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-54m6n"] Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.356120 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-cdj9v" podStartSLOduration=2.601700911 podStartE2EDuration="5.356102675s" podCreationTimestamp="2025-11-24 19:26:56 +0000 UTC" firstStartedPulling="2025-11-24 19:26:57.516360284 +0000 UTC m=+756.038866541" lastFinishedPulling="2025-11-24 19:27:00.270762048 +0000 UTC m=+758.793268305" observedRunningTime="2025-11-24 19:27:01.347453328 +0000 UTC m=+759.869959585" watchObservedRunningTime="2025-11-24 19:27:01.356102675 +0000 UTC m=+759.878608922" Nov 24 19:27:01 crc kubenswrapper[5035]: I1124 19:27:01.896384 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bb98bbd74-bml85" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.064626 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-86b7956b78-crbwt"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.065405 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.069339 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.069406 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.069837 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.069975 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.070889 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.072393 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.078496 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.083838 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86b7956b78-crbwt"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.145000 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-client-ca\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.145140 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-config\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.145188 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16127ec2-6a64-4313-ad34-b59efb64dd50-serving-cert\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.145240 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-proxy-ca-bundles\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.145355 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2v7v\" (UniqueName: \"kubernetes.io/projected/16127ec2-6a64-4313-ad34-b59efb64dd50-kube-api-access-h2v7v\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.208047 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="727a0b65-6c8b-477d-8743-e3bd88a7c904" path="/var/lib/kubelet/pods/727a0b65-6c8b-477d-8743-e3bd88a7c904/volumes" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.208721 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d50060f5-1b00-483a-914b-c34277882c04" path="/var/lib/kubelet/pods/d50060f5-1b00-483a-914b-c34277882c04/volumes" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.246709 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2v7v\" (UniqueName: \"kubernetes.io/projected/16127ec2-6a64-4313-ad34-b59efb64dd50-kube-api-access-h2v7v\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.246799 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-client-ca\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.246840 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-config\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.246862 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16127ec2-6a64-4313-ad34-b59efb64dd50-serving-cert\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.246894 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-proxy-ca-bundles\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.247876 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-client-ca\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.248026 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-proxy-ca-bundles\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.249377 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16127ec2-6a64-4313-ad34-b59efb64dd50-config\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.267182 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16127ec2-6a64-4313-ad34-b59efb64dd50-serving-cert\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.274165 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2v7v\" (UniqueName: \"kubernetes.io/projected/16127ec2-6a64-4313-ad34-b59efb64dd50-kube-api-access-h2v7v\") pod \"controller-manager-86b7956b78-crbwt\" (UID: \"16127ec2-6a64-4313-ad34-b59efb64dd50\") " pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.292745 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.293841 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.312443 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.313201 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.327000 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.343129 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.379677 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.391602 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.422043 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vg29s"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.422874 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.451082 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwdt9\" (UniqueName: \"kubernetes.io/projected/02230763-8196-4df6-8d05-0e6940217381-kube-api-access-nwdt9\") pod \"nmstate-metrics-5dcf9c57c5-rg4vj\" (UID: \"02230763-8196-4df6-8d05-0e6940217381\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.451138 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/74f8a61e-f6be-4f79-9eb9-7a196d884e29-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-fxwc7\" (UID: \"74f8a61e-f6be-4f79-9eb9-7a196d884e29\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.451169 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5cxv\" (UniqueName: \"kubernetes.io/projected/74f8a61e-f6be-4f79-9eb9-7a196d884e29-kube-api-access-m5cxv\") pod \"nmstate-webhook-6b89b748d8-fxwc7\" (UID: \"74f8a61e-f6be-4f79-9eb9-7a196d884e29\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.480839 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.481956 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.484492 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.484527 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-hdnvz" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.484496 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.493775 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.553908 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b9793c81-a455-4d46-bc0a-768b57675a89-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.553971 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b9793c81-a455-4d46-bc0a-768b57675a89-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554049 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwdt9\" (UniqueName: \"kubernetes.io/projected/02230763-8196-4df6-8d05-0e6940217381-kube-api-access-nwdt9\") pod \"nmstate-metrics-5dcf9c57c5-rg4vj\" (UID: \"02230763-8196-4df6-8d05-0e6940217381\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554074 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/74f8a61e-f6be-4f79-9eb9-7a196d884e29-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-fxwc7\" (UID: \"74f8a61e-f6be-4f79-9eb9-7a196d884e29\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554106 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5cxv\" (UniqueName: \"kubernetes.io/projected/74f8a61e-f6be-4f79-9eb9-7a196d884e29-kube-api-access-m5cxv\") pod \"nmstate-webhook-6b89b748d8-fxwc7\" (UID: \"74f8a61e-f6be-4f79-9eb9-7a196d884e29\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554139 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jkhr\" (UniqueName: \"kubernetes.io/projected/c3173bd9-7b88-45d8-bb49-197c85649bc1-kube-api-access-7jkhr\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554161 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w24xs\" (UniqueName: \"kubernetes.io/projected/b9793c81-a455-4d46-bc0a-768b57675a89-kube-api-access-w24xs\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554201 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-nmstate-lock\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554248 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-ovs-socket\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.554302 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-dbus-socket\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: E1124 19:27:02.556494 5035 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 24 19:27:02 crc kubenswrapper[5035]: E1124 19:27:02.556559 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74f8a61e-f6be-4f79-9eb9-7a196d884e29-tls-key-pair podName:74f8a61e-f6be-4f79-9eb9-7a196d884e29 nodeName:}" failed. No retries permitted until 2025-11-24 19:27:03.056537038 +0000 UTC m=+761.579043295 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/74f8a61e-f6be-4f79-9eb9-7a196d884e29-tls-key-pair") pod "nmstate-webhook-6b89b748d8-fxwc7" (UID: "74f8a61e-f6be-4f79-9eb9-7a196d884e29") : secret "openshift-nmstate-webhook" not found Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.583857 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwdt9\" (UniqueName: \"kubernetes.io/projected/02230763-8196-4df6-8d05-0e6940217381-kube-api-access-nwdt9\") pod \"nmstate-metrics-5dcf9c57c5-rg4vj\" (UID: \"02230763-8196-4df6-8d05-0e6940217381\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.595592 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5cxv\" (UniqueName: \"kubernetes.io/projected/74f8a61e-f6be-4f79-9eb9-7a196d884e29-kube-api-access-m5cxv\") pod \"nmstate-webhook-6b89b748d8-fxwc7\" (UID: \"74f8a61e-f6be-4f79-9eb9-7a196d884e29\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.609359 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657305 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b9793c81-a455-4d46-bc0a-768b57675a89-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657348 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b9793c81-a455-4d46-bc0a-768b57675a89-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657397 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jkhr\" (UniqueName: \"kubernetes.io/projected/c3173bd9-7b88-45d8-bb49-197c85649bc1-kube-api-access-7jkhr\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657412 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w24xs\" (UniqueName: \"kubernetes.io/projected/b9793c81-a455-4d46-bc0a-768b57675a89-kube-api-access-w24xs\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657434 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-nmstate-lock\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657455 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-ovs-socket\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657477 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-dbus-socket\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657691 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-dbus-socket\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.657936 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-nmstate-lock\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.658337 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/c3173bd9-7b88-45d8-bb49-197c85649bc1-ovs-socket\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.660731 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b9793c81-a455-4d46-bc0a-768b57675a89-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.661016 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b9793c81-a455-4d46-bc0a-768b57675a89-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.666879 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-65c4f977c4-kxjfn"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.667548 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.679775 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w24xs\" (UniqueName: \"kubernetes.io/projected/b9793c81-a455-4d46-bc0a-768b57675a89-kube-api-access-w24xs\") pod \"nmstate-console-plugin-5874bd7bc5-gz87f\" (UID: \"b9793c81-a455-4d46-bc0a-768b57675a89\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.681189 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jkhr\" (UniqueName: \"kubernetes.io/projected/c3173bd9-7b88-45d8-bb49-197c85649bc1-kube-api-access-7jkhr\") pod \"nmstate-handler-vg29s\" (UID: \"c3173bd9-7b88-45d8-bb49-197c85649bc1\") " pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.682996 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86b7956b78-crbwt"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.687525 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-65c4f977c4-kxjfn"] Nov 24 19:27:02 crc kubenswrapper[5035]: W1124 19:27:02.693315 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16127ec2_6a64_4313_ad34_b59efb64dd50.slice/crio-02633398f868f80bd5a8de448d8bbce77ef960bffa6737575b73dbfa0427fad6 WatchSource:0}: Error finding container 02633398f868f80bd5a8de448d8bbce77ef960bffa6737575b73dbfa0427fad6: Status 404 returned error can't find the container with id 02633398f868f80bd5a8de448d8bbce77ef960bffa6737575b73dbfa0427fad6 Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.758651 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-oauth-config\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.758962 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-trusted-ca-bundle\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.758981 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/8d8a8b1d-ca72-4024-a64b-31b9a2434502-kube-api-access-6st6l\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.758997 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-service-ca\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.759016 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-config\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.759050 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-serving-cert\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.759090 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-oauth-serving-cert\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.761704 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.804135 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860029 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-oauth-config\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860099 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-trusted-ca-bundle\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860121 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/8d8a8b1d-ca72-4024-a64b-31b9a2434502-kube-api-access-6st6l\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860139 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-service-ca\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860164 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-config\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860213 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-serving-cert\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860257 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-oauth-serving-cert\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.860329 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj"] Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.861404 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-oauth-serving-cert\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.861556 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-config\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.861604 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-service-ca\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: W1124 19:27:02.861795 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02230763_8196_4df6_8d05_0e6940217381.slice/crio-4bddebbf105d73635f1a7cc886e9359c72383c768f240c7af08b8a9e08dfba15 WatchSource:0}: Error finding container 4bddebbf105d73635f1a7cc886e9359c72383c768f240c7af08b8a9e08dfba15: Status 404 returned error can't find the container with id 4bddebbf105d73635f1a7cc886e9359c72383c768f240c7af08b8a9e08dfba15 Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.862087 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8d8a8b1d-ca72-4024-a64b-31b9a2434502-trusted-ca-bundle\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.864554 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-serving-cert\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.864799 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8d8a8b1d-ca72-4024-a64b-31b9a2434502-console-oauth-config\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.877733 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6st6l\" (UniqueName: \"kubernetes.io/projected/8d8a8b1d-ca72-4024-a64b-31b9a2434502-kube-api-access-6st6l\") pod \"console-65c4f977c4-kxjfn\" (UID: \"8d8a8b1d-ca72-4024-a64b-31b9a2434502\") " pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:02 crc kubenswrapper[5035]: I1124 19:27:02.990935 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.004318 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f"] Nov 24 19:27:03 crc kubenswrapper[5035]: W1124 19:27:03.008767 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9793c81_a455_4d46_bc0a_768b57675a89.slice/crio-04c036d9f66cb69736c451c65cce94541ac7e08fc2a3dfca7c6a73c368b53549 WatchSource:0}: Error finding container 04c036d9f66cb69736c451c65cce94541ac7e08fc2a3dfca7c6a73c368b53549: Status 404 returned error can't find the container with id 04c036d9f66cb69736c451c65cce94541ac7e08fc2a3dfca7c6a73c368b53549 Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.063236 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/74f8a61e-f6be-4f79-9eb9-7a196d884e29-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-fxwc7\" (UID: \"74f8a61e-f6be-4f79-9eb9-7a196d884e29\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.067905 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/74f8a61e-f6be-4f79-9eb9-7a196d884e29-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-fxwc7\" (UID: \"74f8a61e-f6be-4f79-9eb9-7a196d884e29\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.282318 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.299068 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" event={"ID":"02230763-8196-4df6-8d05-0e6940217381","Type":"ContainerStarted","Data":"4bddebbf105d73635f1a7cc886e9359c72383c768f240c7af08b8a9e08dfba15"} Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.306852 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" event={"ID":"b9793c81-a455-4d46-bc0a-768b57675a89","Type":"ContainerStarted","Data":"04c036d9f66cb69736c451c65cce94541ac7e08fc2a3dfca7c6a73c368b53549"} Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.309164 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vg29s" event={"ID":"c3173bd9-7b88-45d8-bb49-197c85649bc1","Type":"ContainerStarted","Data":"021a0429290afcee6405d34913dba063e28dacc93baf52d806ab53dca517d679"} Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.310707 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" event={"ID":"16127ec2-6a64-4313-ad34-b59efb64dd50","Type":"ContainerStarted","Data":"c2a81d9b69cb129dd6cef75cf7c77142bc705e6309d372cf2e22300efc94ca59"} Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.310729 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" event={"ID":"16127ec2-6a64-4313-ad34-b59efb64dd50","Type":"ContainerStarted","Data":"02633398f868f80bd5a8de448d8bbce77ef960bffa6737575b73dbfa0427fad6"} Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.326929 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-65c4f977c4-kxjfn"] Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.335144 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" podStartSLOduration=3.335129182 podStartE2EDuration="3.335129182s" podCreationTimestamp="2025-11-24 19:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:27:03.333611131 +0000 UTC m=+761.856117388" watchObservedRunningTime="2025-11-24 19:27:03.335129182 +0000 UTC m=+761.857635439" Nov 24 19:27:03 crc kubenswrapper[5035]: I1124 19:27:03.546498 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7"] Nov 24 19:27:03 crc kubenswrapper[5035]: W1124 19:27:03.559641 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f8a61e_f6be_4f79_9eb9_7a196d884e29.slice/crio-b1e8216f817e4ae48b37f6491dfe69098a1ef5cf8ddf137a9ab44308bee5ee3c WatchSource:0}: Error finding container b1e8216f817e4ae48b37f6491dfe69098a1ef5cf8ddf137a9ab44308bee5ee3c: Status 404 returned error can't find the container with id b1e8216f817e4ae48b37f6491dfe69098a1ef5cf8ddf137a9ab44308bee5ee3c Nov 24 19:27:04 crc kubenswrapper[5035]: I1124 19:27:04.322367 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-65c4f977c4-kxjfn" event={"ID":"8d8a8b1d-ca72-4024-a64b-31b9a2434502","Type":"ContainerStarted","Data":"f28cff728a0eeac60243cea9d78d1d5807438206efb0ac7e12e04ead7ee168ad"} Nov 24 19:27:04 crc kubenswrapper[5035]: I1124 19:27:04.322875 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-65c4f977c4-kxjfn" event={"ID":"8d8a8b1d-ca72-4024-a64b-31b9a2434502","Type":"ContainerStarted","Data":"5768370f4cb68396c758110d241a8ed60fd4fbc8755f3118b5f29563b6928462"} Nov 24 19:27:04 crc kubenswrapper[5035]: I1124 19:27:04.323855 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" event={"ID":"74f8a61e-f6be-4f79-9eb9-7a196d884e29","Type":"ContainerStarted","Data":"b1e8216f817e4ae48b37f6491dfe69098a1ef5cf8ddf137a9ab44308bee5ee3c"} Nov 24 19:27:04 crc kubenswrapper[5035]: I1124 19:27:04.324216 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:04 crc kubenswrapper[5035]: I1124 19:27:04.330484 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86b7956b78-crbwt" Nov 24 19:27:04 crc kubenswrapper[5035]: I1124 19:27:04.342504 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-65c4f977c4-kxjfn" podStartSLOduration=2.342483601 podStartE2EDuration="2.342483601s" podCreationTimestamp="2025-11-24 19:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:27:04.336878677 +0000 UTC m=+762.859384934" watchObservedRunningTime="2025-11-24 19:27:04.342483601 +0000 UTC m=+762.864989858" Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.334356 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vg29s" event={"ID":"c3173bd9-7b88-45d8-bb49-197c85649bc1","Type":"ContainerStarted","Data":"700b629117b89eab8ee10dfa2d4ffe1d8f31b03f4500ea23d32aa2bd4fc825e8"} Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.334860 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.336112 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" event={"ID":"02230763-8196-4df6-8d05-0e6940217381","Type":"ContainerStarted","Data":"1ecb653d2377d52fe774dd9a3a8286f6c403dc6f07a4af682a21637ac12ad1e2"} Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.337456 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" event={"ID":"74f8a61e-f6be-4f79-9eb9-7a196d884e29","Type":"ContainerStarted","Data":"131ce6ea84a9de7cf158e5b58a58cf24c78b58b3c5a5e8567235f2243a0c898b"} Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.337561 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.339602 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" event={"ID":"b9793c81-a455-4d46-bc0a-768b57675a89","Type":"ContainerStarted","Data":"eba74c535449b3d43a9f0a9ab79e197284bd4b64dcd8d36830e30cd5a64ca962"} Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.381017 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vg29s" podStartSLOduration=1.225357592 podStartE2EDuration="4.38099575s" podCreationTimestamp="2025-11-24 19:27:02 +0000 UTC" firstStartedPulling="2025-11-24 19:27:02.789780216 +0000 UTC m=+761.312286473" lastFinishedPulling="2025-11-24 19:27:05.945418374 +0000 UTC m=+764.467924631" observedRunningTime="2025-11-24 19:27:06.356378464 +0000 UTC m=+764.878884731" watchObservedRunningTime="2025-11-24 19:27:06.38099575 +0000 UTC m=+764.903502017" Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.382409 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-gz87f" podStartSLOduration=1.483729498 podStartE2EDuration="4.382398388s" podCreationTimestamp="2025-11-24 19:27:02 +0000 UTC" firstStartedPulling="2025-11-24 19:27:03.011444225 +0000 UTC m=+761.533950482" lastFinishedPulling="2025-11-24 19:27:05.910113115 +0000 UTC m=+764.432619372" observedRunningTime="2025-11-24 19:27:06.369844264 +0000 UTC m=+764.892350551" watchObservedRunningTime="2025-11-24 19:27:06.382398388 +0000 UTC m=+764.904904655" Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.401619 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" podStartSLOduration=1.99718469 podStartE2EDuration="4.401596194s" podCreationTimestamp="2025-11-24 19:27:02 +0000 UTC" firstStartedPulling="2025-11-24 19:27:03.564218376 +0000 UTC m=+762.086724633" lastFinishedPulling="2025-11-24 19:27:05.96862988 +0000 UTC m=+764.491136137" observedRunningTime="2025-11-24 19:27:06.398626783 +0000 UTC m=+764.921133050" watchObservedRunningTime="2025-11-24 19:27:06.401596194 +0000 UTC m=+764.924102471" Nov 24 19:27:06 crc kubenswrapper[5035]: I1124 19:27:06.465324 5035 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 19:27:09 crc kubenswrapper[5035]: I1124 19:27:09.361543 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" event={"ID":"02230763-8196-4df6-8d05-0e6940217381","Type":"ContainerStarted","Data":"f7f1b17080bb64ca1222cde4ee6f90697b4f9abafc4e02ea165ba22ba93abbe8"} Nov 24 19:27:09 crc kubenswrapper[5035]: I1124 19:27:09.385471 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-rg4vj" podStartSLOduration=1.54393489 podStartE2EDuration="7.385454831s" podCreationTimestamp="2025-11-24 19:27:02 +0000 UTC" firstStartedPulling="2025-11-24 19:27:02.865443201 +0000 UTC m=+761.387949458" lastFinishedPulling="2025-11-24 19:27:08.706963152 +0000 UTC m=+767.229469399" observedRunningTime="2025-11-24 19:27:09.384369161 +0000 UTC m=+767.906875448" watchObservedRunningTime="2025-11-24 19:27:09.385454831 +0000 UTC m=+767.907961118" Nov 24 19:27:12 crc kubenswrapper[5035]: I1124 19:27:12.789884 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vg29s" Nov 24 19:27:12 crc kubenswrapper[5035]: I1124 19:27:12.991861 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:12 crc kubenswrapper[5035]: I1124 19:27:12.992146 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:13 crc kubenswrapper[5035]: I1124 19:27:13.000902 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:13 crc kubenswrapper[5035]: I1124 19:27:13.390475 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-65c4f977c4-kxjfn" Nov 24 19:27:13 crc kubenswrapper[5035]: I1124 19:27:13.447584 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-58jzg"] Nov 24 19:27:15 crc kubenswrapper[5035]: I1124 19:27:15.234740 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:27:15 crc kubenswrapper[5035]: I1124 19:27:15.234828 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:27:23 crc kubenswrapper[5035]: I1124 19:27:23.289758 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-fxwc7" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.836188 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k25vz"] Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.839635 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.860790 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k25vz"] Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.886945 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f9q8\" (UniqueName: \"kubernetes.io/projected/c1e46744-dff9-496c-aace-da5c5dad8776-kube-api-access-6f9q8\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.887035 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-utilities\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.887140 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-catalog-content\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.990162 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-utilities\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.990275 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-catalog-content\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.990321 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f9q8\" (UniqueName: \"kubernetes.io/projected/c1e46744-dff9-496c-aace-da5c5dad8776-kube-api-access-6f9q8\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.990787 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-utilities\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:32 crc kubenswrapper[5035]: I1124 19:27:32.990838 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-catalog-content\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:33 crc kubenswrapper[5035]: I1124 19:27:33.012605 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f9q8\" (UniqueName: \"kubernetes.io/projected/c1e46744-dff9-496c-aace-da5c5dad8776-kube-api-access-6f9q8\") pod \"redhat-operators-k25vz\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:33 crc kubenswrapper[5035]: I1124 19:27:33.182268 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:33 crc kubenswrapper[5035]: I1124 19:27:33.592634 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k25vz"] Nov 24 19:27:34 crc kubenswrapper[5035]: I1124 19:27:34.530687 5035 generic.go:334] "Generic (PLEG): container finished" podID="c1e46744-dff9-496c-aace-da5c5dad8776" containerID="68b3a6cc6938f33b7227a2fb14da56aae80e4dbbecfa48e11997c00ee2f665d1" exitCode=0 Nov 24 19:27:34 crc kubenswrapper[5035]: I1124 19:27:34.530751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k25vz" event={"ID":"c1e46744-dff9-496c-aace-da5c5dad8776","Type":"ContainerDied","Data":"68b3a6cc6938f33b7227a2fb14da56aae80e4dbbecfa48e11997c00ee2f665d1"} Nov 24 19:27:34 crc kubenswrapper[5035]: I1124 19:27:34.531211 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k25vz" event={"ID":"c1e46744-dff9-496c-aace-da5c5dad8776","Type":"ContainerStarted","Data":"7c8f506443a8b8e0397180bd299c0e5fd622754adb7d399bfdf504c996991d89"} Nov 24 19:27:35 crc kubenswrapper[5035]: I1124 19:27:35.537152 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k25vz" event={"ID":"c1e46744-dff9-496c-aace-da5c5dad8776","Type":"ContainerStarted","Data":"25fdf1c568885d39d7388d62a1afdf47074284e2e33b656a1226ce5184aa0b58"} Nov 24 19:27:36 crc kubenswrapper[5035]: I1124 19:27:36.546405 5035 generic.go:334] "Generic (PLEG): container finished" podID="c1e46744-dff9-496c-aace-da5c5dad8776" containerID="25fdf1c568885d39d7388d62a1afdf47074284e2e33b656a1226ce5184aa0b58" exitCode=0 Nov 24 19:27:36 crc kubenswrapper[5035]: I1124 19:27:36.546811 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k25vz" event={"ID":"c1e46744-dff9-496c-aace-da5c5dad8776","Type":"ContainerDied","Data":"25fdf1c568885d39d7388d62a1afdf47074284e2e33b656a1226ce5184aa0b58"} Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.556045 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k25vz" event={"ID":"c1e46744-dff9-496c-aace-da5c5dad8776","Type":"ContainerStarted","Data":"c15928ef327e943fb637ea6df8c44f31b133bb61f7ff8634c4a3f60e5c87774a"} Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.574892 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k25vz" podStartSLOduration=2.9792114659999998 podStartE2EDuration="5.574873816s" podCreationTimestamp="2025-11-24 19:27:32 +0000 UTC" firstStartedPulling="2025-11-24 19:27:34.535520187 +0000 UTC m=+793.058026444" lastFinishedPulling="2025-11-24 19:27:37.131182537 +0000 UTC m=+795.653688794" observedRunningTime="2025-11-24 19:27:37.572813609 +0000 UTC m=+796.095319866" watchObservedRunningTime="2025-11-24 19:27:37.574873816 +0000 UTC m=+796.097380073" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.671943 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4"] Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.673200 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.675949 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.686108 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4"] Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.749807 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6l4g\" (UniqueName: \"kubernetes.io/projected/a606927e-ed47-44e2-a593-8dc09c393d71-kube-api-access-z6l4g\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.750127 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.750237 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.851923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.852040 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.852133 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6l4g\" (UniqueName: \"kubernetes.io/projected/a606927e-ed47-44e2-a593-8dc09c393d71-kube-api-access-z6l4g\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.852845 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.853310 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.877387 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6l4g\" (UniqueName: \"kubernetes.io/projected/a606927e-ed47-44e2-a593-8dc09c393d71-kube-api-access-z6l4g\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:37 crc kubenswrapper[5035]: I1124 19:27:37.987976 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:38 crc kubenswrapper[5035]: I1124 19:27:38.428089 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4"] Nov 24 19:27:38 crc kubenswrapper[5035]: W1124 19:27:38.449560 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda606927e_ed47_44e2_a593_8dc09c393d71.slice/crio-66485aace2d0e0d576f457f371e8210b6bc4fc1eb6a6f6de6b42286b0366d6c9 WatchSource:0}: Error finding container 66485aace2d0e0d576f457f371e8210b6bc4fc1eb6a6f6de6b42286b0366d6c9: Status 404 returned error can't find the container with id 66485aace2d0e0d576f457f371e8210b6bc4fc1eb6a6f6de6b42286b0366d6c9 Nov 24 19:27:38 crc kubenswrapper[5035]: I1124 19:27:38.515569 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-58jzg" podUID="2a0d45af-2aba-4830-a849-c2933f94fb83" containerName="console" containerID="cri-o://0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a" gracePeriod=15 Nov 24 19:27:38 crc kubenswrapper[5035]: I1124 19:27:38.561946 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" event={"ID":"a606927e-ed47-44e2-a593-8dc09c393d71","Type":"ContainerStarted","Data":"66485aace2d0e0d576f457f371e8210b6bc4fc1eb6a6f6de6b42286b0366d6c9"} Nov 24 19:27:38 crc kubenswrapper[5035]: I1124 19:27:38.907946 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-58jzg_2a0d45af-2aba-4830-a849-c2933f94fb83/console/0.log" Nov 24 19:27:38 crc kubenswrapper[5035]: I1124 19:27:38.908197 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.096667 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-console-config\") pod \"2a0d45af-2aba-4830-a849-c2933f94fb83\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.096718 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-trusted-ca-bundle\") pod \"2a0d45af-2aba-4830-a849-c2933f94fb83\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.096823 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-serving-cert\") pod \"2a0d45af-2aba-4830-a849-c2933f94fb83\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.096856 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5wjh\" (UniqueName: \"kubernetes.io/projected/2a0d45af-2aba-4830-a849-c2933f94fb83-kube-api-access-z5wjh\") pod \"2a0d45af-2aba-4830-a849-c2933f94fb83\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.096892 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-oauth-config\") pod \"2a0d45af-2aba-4830-a849-c2933f94fb83\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.096918 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-service-ca\") pod \"2a0d45af-2aba-4830-a849-c2933f94fb83\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.096944 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-oauth-serving-cert\") pod \"2a0d45af-2aba-4830-a849-c2933f94fb83\" (UID: \"2a0d45af-2aba-4830-a849-c2933f94fb83\") " Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.097669 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2a0d45af-2aba-4830-a849-c2933f94fb83" (UID: "2a0d45af-2aba-4830-a849-c2933f94fb83"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.097883 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2a0d45af-2aba-4830-a849-c2933f94fb83" (UID: "2a0d45af-2aba-4830-a849-c2933f94fb83"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.098081 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-service-ca" (OuterVolumeSpecName: "service-ca") pod "2a0d45af-2aba-4830-a849-c2933f94fb83" (UID: "2a0d45af-2aba-4830-a849-c2933f94fb83"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.098141 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-console-config" (OuterVolumeSpecName: "console-config") pod "2a0d45af-2aba-4830-a849-c2933f94fb83" (UID: "2a0d45af-2aba-4830-a849-c2933f94fb83"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.102068 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2a0d45af-2aba-4830-a849-c2933f94fb83" (UID: "2a0d45af-2aba-4830-a849-c2933f94fb83"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.102167 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a0d45af-2aba-4830-a849-c2933f94fb83-kube-api-access-z5wjh" (OuterVolumeSpecName: "kube-api-access-z5wjh") pod "2a0d45af-2aba-4830-a849-c2933f94fb83" (UID: "2a0d45af-2aba-4830-a849-c2933f94fb83"). InnerVolumeSpecName "kube-api-access-z5wjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.102359 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2a0d45af-2aba-4830-a849-c2933f94fb83" (UID: "2a0d45af-2aba-4830-a849-c2933f94fb83"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.198665 5035 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.198701 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5wjh\" (UniqueName: \"kubernetes.io/projected/2a0d45af-2aba-4830-a849-c2933f94fb83-kube-api-access-z5wjh\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.198716 5035 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2a0d45af-2aba-4830-a849-c2933f94fb83-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.198732 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.198744 5035 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.198756 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.198768 5035 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2a0d45af-2aba-4830-a849-c2933f94fb83-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.568805 5035 generic.go:334] "Generic (PLEG): container finished" podID="a606927e-ed47-44e2-a593-8dc09c393d71" containerID="60662074fd0e41cb1a0ef2cb734949ebff954caeef66dd9c7264e18f2395310d" exitCode=0 Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.568869 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" event={"ID":"a606927e-ed47-44e2-a593-8dc09c393d71","Type":"ContainerDied","Data":"60662074fd0e41cb1a0ef2cb734949ebff954caeef66dd9c7264e18f2395310d"} Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.573006 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-58jzg_2a0d45af-2aba-4830-a849-c2933f94fb83/console/0.log" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.573090 5035 generic.go:334] "Generic (PLEG): container finished" podID="2a0d45af-2aba-4830-a849-c2933f94fb83" containerID="0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a" exitCode=2 Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.573137 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58jzg" event={"ID":"2a0d45af-2aba-4830-a849-c2933f94fb83","Type":"ContainerDied","Data":"0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a"} Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.573176 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-58jzg" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.573204 5035 scope.go:117] "RemoveContainer" containerID="0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.573184 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-58jzg" event={"ID":"2a0d45af-2aba-4830-a849-c2933f94fb83","Type":"ContainerDied","Data":"71e590e92e30f69e7d0c51233569a82504b33552e6054792779a19474ef433cb"} Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.595399 5035 scope.go:117] "RemoveContainer" containerID="0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a" Nov 24 19:27:39 crc kubenswrapper[5035]: E1124 19:27:39.597636 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a\": container with ID starting with 0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a not found: ID does not exist" containerID="0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.597727 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a"} err="failed to get container status \"0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a\": rpc error: code = NotFound desc = could not find container \"0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a\": container with ID starting with 0d0eff2b43a506c3f1e7f1ff0bfb366c8dded75ef8cab6a7c088e0edf848bb4a not found: ID does not exist" Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.619081 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-58jzg"] Nov 24 19:27:39 crc kubenswrapper[5035]: I1124 19:27:39.625882 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-58jzg"] Nov 24 19:27:40 crc kubenswrapper[5035]: I1124 19:27:40.213315 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a0d45af-2aba-4830-a849-c2933f94fb83" path="/var/lib/kubelet/pods/2a0d45af-2aba-4830-a849-c2933f94fb83/volumes" Nov 24 19:27:41 crc kubenswrapper[5035]: I1124 19:27:41.591490 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" event={"ID":"a606927e-ed47-44e2-a593-8dc09c393d71","Type":"ContainerStarted","Data":"6f9abc097fe75ac1db6ecadf5ab9ad7690b76b00af1ade95db531eb5050f94ea"} Nov 24 19:27:43 crc kubenswrapper[5035]: I1124 19:27:43.182641 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:43 crc kubenswrapper[5035]: I1124 19:27:43.183074 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:43 crc kubenswrapper[5035]: I1124 19:27:43.244460 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:43 crc kubenswrapper[5035]: I1124 19:27:43.602408 5035 generic.go:334] "Generic (PLEG): container finished" podID="a606927e-ed47-44e2-a593-8dc09c393d71" containerID="6f9abc097fe75ac1db6ecadf5ab9ad7690b76b00af1ade95db531eb5050f94ea" exitCode=0 Nov 24 19:27:43 crc kubenswrapper[5035]: I1124 19:27:43.602506 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" event={"ID":"a606927e-ed47-44e2-a593-8dc09c393d71","Type":"ContainerDied","Data":"6f9abc097fe75ac1db6ecadf5ab9ad7690b76b00af1ade95db531eb5050f94ea"} Nov 24 19:27:43 crc kubenswrapper[5035]: I1124 19:27:43.647783 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:44 crc kubenswrapper[5035]: I1124 19:27:44.610948 5035 generic.go:334] "Generic (PLEG): container finished" podID="a606927e-ed47-44e2-a593-8dc09c393d71" containerID="bcd1c6c8d97cac7328b0caa1c86659e7e527c677b58976c061b039ee503b4696" exitCode=0 Nov 24 19:27:44 crc kubenswrapper[5035]: I1124 19:27:44.611032 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" event={"ID":"a606927e-ed47-44e2-a593-8dc09c393d71","Type":"ContainerDied","Data":"bcd1c6c8d97cac7328b0caa1c86659e7e527c677b58976c061b039ee503b4696"} Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.234606 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.234984 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.235058 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.235987 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"88531e6ce05adbac2436f424c2cf6aa59384e9fc263400422a229df68ffdfb39"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.236094 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://88531e6ce05adbac2436f424c2cf6aa59384e9fc263400422a229df68ffdfb39" gracePeriod=600 Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.620802 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k25vz"] Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.630647 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="88531e6ce05adbac2436f424c2cf6aa59384e9fc263400422a229df68ffdfb39" exitCode=0 Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.630901 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"88531e6ce05adbac2436f424c2cf6aa59384e9fc263400422a229df68ffdfb39"} Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.630946 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"7f8c76c75a096d1e632ca8dbc062a2c86c922da2cd856db7b41a1460c00d789d"} Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.630969 5035 scope.go:117] "RemoveContainer" containerID="b978ecf55cf22195f7615a97b366d61585baa053ee9c7eac1b8d508bebaa7f5e" Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.631139 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k25vz" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="registry-server" containerID="cri-o://c15928ef327e943fb637ea6df8c44f31b133bb61f7ff8634c4a3f60e5c87774a" gracePeriod=2 Nov 24 19:27:45 crc kubenswrapper[5035]: I1124 19:27:45.895389 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.085647 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6l4g\" (UniqueName: \"kubernetes.io/projected/a606927e-ed47-44e2-a593-8dc09c393d71-kube-api-access-z6l4g\") pod \"a606927e-ed47-44e2-a593-8dc09c393d71\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.085853 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-bundle\") pod \"a606927e-ed47-44e2-a593-8dc09c393d71\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.085934 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-util\") pod \"a606927e-ed47-44e2-a593-8dc09c393d71\" (UID: \"a606927e-ed47-44e2-a593-8dc09c393d71\") " Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.087461 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-bundle" (OuterVolumeSpecName: "bundle") pod "a606927e-ed47-44e2-a593-8dc09c393d71" (UID: "a606927e-ed47-44e2-a593-8dc09c393d71"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.094089 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a606927e-ed47-44e2-a593-8dc09c393d71-kube-api-access-z6l4g" (OuterVolumeSpecName: "kube-api-access-z6l4g") pod "a606927e-ed47-44e2-a593-8dc09c393d71" (UID: "a606927e-ed47-44e2-a593-8dc09c393d71"). InnerVolumeSpecName "kube-api-access-z6l4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.096986 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-util" (OuterVolumeSpecName: "util") pod "a606927e-ed47-44e2-a593-8dc09c393d71" (UID: "a606927e-ed47-44e2-a593-8dc09c393d71"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.187645 5035 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.187679 5035 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a606927e-ed47-44e2-a593-8dc09c393d71-util\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.187689 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6l4g\" (UniqueName: \"kubernetes.io/projected/a606927e-ed47-44e2-a593-8dc09c393d71-kube-api-access-z6l4g\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.641765 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" event={"ID":"a606927e-ed47-44e2-a593-8dc09c393d71","Type":"ContainerDied","Data":"66485aace2d0e0d576f457f371e8210b6bc4fc1eb6a6f6de6b42286b0366d6c9"} Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.641809 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66485aace2d0e0d576f457f371e8210b6bc4fc1eb6a6f6de6b42286b0366d6c9" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.641889 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4" Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.649784 5035 generic.go:334] "Generic (PLEG): container finished" podID="c1e46744-dff9-496c-aace-da5c5dad8776" containerID="c15928ef327e943fb637ea6df8c44f31b133bb61f7ff8634c4a3f60e5c87774a" exitCode=0 Nov 24 19:27:46 crc kubenswrapper[5035]: I1124 19:27:46.649853 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k25vz" event={"ID":"c1e46744-dff9-496c-aace-da5c5dad8776","Type":"ContainerDied","Data":"c15928ef327e943fb637ea6df8c44f31b133bb61f7ff8634c4a3f60e5c87774a"} Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.114277 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.300849 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f9q8\" (UniqueName: \"kubernetes.io/projected/c1e46744-dff9-496c-aace-da5c5dad8776-kube-api-access-6f9q8\") pod \"c1e46744-dff9-496c-aace-da5c5dad8776\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.301010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-catalog-content\") pod \"c1e46744-dff9-496c-aace-da5c5dad8776\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.301122 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-utilities\") pod \"c1e46744-dff9-496c-aace-da5c5dad8776\" (UID: \"c1e46744-dff9-496c-aace-da5c5dad8776\") " Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.302308 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-utilities" (OuterVolumeSpecName: "utilities") pod "c1e46744-dff9-496c-aace-da5c5dad8776" (UID: "c1e46744-dff9-496c-aace-da5c5dad8776"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.307202 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1e46744-dff9-496c-aace-da5c5dad8776-kube-api-access-6f9q8" (OuterVolumeSpecName: "kube-api-access-6f9q8") pod "c1e46744-dff9-496c-aace-da5c5dad8776" (UID: "c1e46744-dff9-496c-aace-da5c5dad8776"). InnerVolumeSpecName "kube-api-access-6f9q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.402548 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f9q8\" (UniqueName: \"kubernetes.io/projected/c1e46744-dff9-496c-aace-da5c5dad8776-kube-api-access-6f9q8\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.402601 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.413417 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1e46744-dff9-496c-aace-da5c5dad8776" (UID: "c1e46744-dff9-496c-aace-da5c5dad8776"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.504341 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1e46744-dff9-496c-aace-da5c5dad8776-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.661244 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k25vz" event={"ID":"c1e46744-dff9-496c-aace-da5c5dad8776","Type":"ContainerDied","Data":"7c8f506443a8b8e0397180bd299c0e5fd622754adb7d399bfdf504c996991d89"} Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.661320 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k25vz" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.661346 5035 scope.go:117] "RemoveContainer" containerID="c15928ef327e943fb637ea6df8c44f31b133bb61f7ff8634c4a3f60e5c87774a" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.688450 5035 scope.go:117] "RemoveContainer" containerID="25fdf1c568885d39d7388d62a1afdf47074284e2e33b656a1226ce5184aa0b58" Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.701115 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k25vz"] Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.705336 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k25vz"] Nov 24 19:27:47 crc kubenswrapper[5035]: I1124 19:27:47.726057 5035 scope.go:117] "RemoveContainer" containerID="68b3a6cc6938f33b7227a2fb14da56aae80e4dbbecfa48e11997c00ee2f665d1" Nov 24 19:27:48 crc kubenswrapper[5035]: I1124 19:27:48.206426 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" path="/var/lib/kubelet/pods/c1e46744-dff9-496c-aace-da5c5dad8776/volumes" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.620900 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8chpn"] Nov 24 19:27:52 crc kubenswrapper[5035]: E1124 19:27:52.621833 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0d45af-2aba-4830-a849-c2933f94fb83" containerName="console" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.621854 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0d45af-2aba-4830-a849-c2933f94fb83" containerName="console" Nov 24 19:27:52 crc kubenswrapper[5035]: E1124 19:27:52.621884 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a606927e-ed47-44e2-a593-8dc09c393d71" containerName="pull" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.621895 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="a606927e-ed47-44e2-a593-8dc09c393d71" containerName="pull" Nov 24 19:27:52 crc kubenswrapper[5035]: E1124 19:27:52.621912 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="extract-utilities" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.621925 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="extract-utilities" Nov 24 19:27:52 crc kubenswrapper[5035]: E1124 19:27:52.621941 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a606927e-ed47-44e2-a593-8dc09c393d71" containerName="util" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.621952 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="a606927e-ed47-44e2-a593-8dc09c393d71" containerName="util" Nov 24 19:27:52 crc kubenswrapper[5035]: E1124 19:27:52.621969 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a606927e-ed47-44e2-a593-8dc09c393d71" containerName="extract" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.621979 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="a606927e-ed47-44e2-a593-8dc09c393d71" containerName="extract" Nov 24 19:27:52 crc kubenswrapper[5035]: E1124 19:27:52.621995 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="extract-content" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.622005 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="extract-content" Nov 24 19:27:52 crc kubenswrapper[5035]: E1124 19:27:52.622022 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="registry-server" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.622032 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="registry-server" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.622188 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="a606927e-ed47-44e2-a593-8dc09c393d71" containerName="extract" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.622209 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e46744-dff9-496c-aace-da5c5dad8776" containerName="registry-server" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.622222 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0d45af-2aba-4830-a849-c2933f94fb83" containerName="console" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.623326 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.632165 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8chpn"] Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.771207 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-utilities\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.771423 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-catalog-content\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.771480 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hszv4\" (UniqueName: \"kubernetes.io/projected/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-kube-api-access-hszv4\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.872811 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-utilities\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.872896 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-catalog-content\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.872932 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hszv4\" (UniqueName: \"kubernetes.io/projected/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-kube-api-access-hszv4\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.873509 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-utilities\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.873533 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-catalog-content\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.906882 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hszv4\" (UniqueName: \"kubernetes.io/projected/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-kube-api-access-hszv4\") pod \"redhat-marketplace-8chpn\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:52 crc kubenswrapper[5035]: I1124 19:27:52.943098 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:27:53 crc kubenswrapper[5035]: I1124 19:27:53.382067 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8chpn"] Nov 24 19:27:53 crc kubenswrapper[5035]: I1124 19:27:53.695814 5035 generic.go:334] "Generic (PLEG): container finished" podID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerID="481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e" exitCode=0 Nov 24 19:27:53 crc kubenswrapper[5035]: I1124 19:27:53.695855 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8chpn" event={"ID":"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2","Type":"ContainerDied","Data":"481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e"} Nov 24 19:27:53 crc kubenswrapper[5035]: I1124 19:27:53.695877 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8chpn" event={"ID":"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2","Type":"ContainerStarted","Data":"0168cb1f3d8b4286fad32c67f03dede3f62dea9bc2d5869aa252b5dc66cabbb0"} Nov 24 19:27:55 crc kubenswrapper[5035]: I1124 19:27:55.710376 5035 generic.go:334] "Generic (PLEG): container finished" podID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerID="1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3" exitCode=0 Nov 24 19:27:55 crc kubenswrapper[5035]: I1124 19:27:55.710437 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8chpn" event={"ID":"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2","Type":"ContainerDied","Data":"1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3"} Nov 24 19:27:56 crc kubenswrapper[5035]: I1124 19:27:56.718925 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8chpn" event={"ID":"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2","Type":"ContainerStarted","Data":"ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3"} Nov 24 19:27:56 crc kubenswrapper[5035]: I1124 19:27:56.738836 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8chpn" podStartSLOduration=2.076064123 podStartE2EDuration="4.738813143s" podCreationTimestamp="2025-11-24 19:27:52 +0000 UTC" firstStartedPulling="2025-11-24 19:27:53.69781972 +0000 UTC m=+812.220325987" lastFinishedPulling="2025-11-24 19:27:56.36056875 +0000 UTC m=+814.883075007" observedRunningTime="2025-11-24 19:27:56.736015327 +0000 UTC m=+815.258521584" watchObservedRunningTime="2025-11-24 19:27:56.738813143 +0000 UTC m=+815.261319420" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.314627 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk"] Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.315857 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.318446 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.318478 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.318646 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.320417 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7gpwm" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.322644 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.389438 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk"] Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.436573 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxjt9\" (UniqueName: \"kubernetes.io/projected/b782ab87-baed-4a09-b133-affcf2efd744-kube-api-access-nxjt9\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.436629 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b782ab87-baed-4a09-b133-affcf2efd744-webhook-cert\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.436665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b782ab87-baed-4a09-b133-affcf2efd744-apiservice-cert\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.538344 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b782ab87-baed-4a09-b133-affcf2efd744-webhook-cert\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.538420 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b782ab87-baed-4a09-b133-affcf2efd744-apiservice-cert\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.538556 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxjt9\" (UniqueName: \"kubernetes.io/projected/b782ab87-baed-4a09-b133-affcf2efd744-kube-api-access-nxjt9\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.544071 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b782ab87-baed-4a09-b133-affcf2efd744-webhook-cert\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.544838 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b782ab87-baed-4a09-b133-affcf2efd744-apiservice-cert\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.556782 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxjt9\" (UniqueName: \"kubernetes.io/projected/b782ab87-baed-4a09-b133-affcf2efd744-kube-api-access-nxjt9\") pod \"metallb-operator-controller-manager-6648cb8994-hqcpk\" (UID: \"b782ab87-baed-4a09-b133-affcf2efd744\") " pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.576538 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t"] Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.577176 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.579425 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.579635 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.579878 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5gcwz" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.596481 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t"] Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.630002 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.740942 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fg2h\" (UniqueName: \"kubernetes.io/projected/60ba6695-0ae0-4dce-b6de-ec96a4723d07-kube-api-access-6fg2h\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.741311 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60ba6695-0ae0-4dce-b6de-ec96a4723d07-webhook-cert\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.741405 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60ba6695-0ae0-4dce-b6de-ec96a4723d07-apiservice-cert\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.842461 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60ba6695-0ae0-4dce-b6de-ec96a4723d07-webhook-cert\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.842567 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60ba6695-0ae0-4dce-b6de-ec96a4723d07-apiservice-cert\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.842613 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fg2h\" (UniqueName: \"kubernetes.io/projected/60ba6695-0ae0-4dce-b6de-ec96a4723d07-kube-api-access-6fg2h\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.850500 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60ba6695-0ae0-4dce-b6de-ec96a4723d07-webhook-cert\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.863937 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60ba6695-0ae0-4dce-b6de-ec96a4723d07-apiservice-cert\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.868492 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fg2h\" (UniqueName: \"kubernetes.io/projected/60ba6695-0ae0-4dce-b6de-ec96a4723d07-kube-api-access-6fg2h\") pod \"metallb-operator-webhook-server-6fbc66c766-4477t\" (UID: \"60ba6695-0ae0-4dce-b6de-ec96a4723d07\") " pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:57 crc kubenswrapper[5035]: I1124 19:27:57.912572 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:27:58 crc kubenswrapper[5035]: I1124 19:27:58.105765 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk"] Nov 24 19:27:58 crc kubenswrapper[5035]: W1124 19:27:58.122362 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb782ab87_baed_4a09_b133_affcf2efd744.slice/crio-b26cfdf80c405eacc941bed3453dea6a10b4f776505d1ff2b679f377ac15d3b0 WatchSource:0}: Error finding container b26cfdf80c405eacc941bed3453dea6a10b4f776505d1ff2b679f377ac15d3b0: Status 404 returned error can't find the container with id b26cfdf80c405eacc941bed3453dea6a10b4f776505d1ff2b679f377ac15d3b0 Nov 24 19:27:58 crc kubenswrapper[5035]: I1124 19:27:58.393210 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t"] Nov 24 19:27:58 crc kubenswrapper[5035]: W1124 19:27:58.400998 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60ba6695_0ae0_4dce_b6de_ec96a4723d07.slice/crio-c87f9d9af6ddab15e67e3d3ca25cd854d16167fa7a67c91d24dadb8619c4543e WatchSource:0}: Error finding container c87f9d9af6ddab15e67e3d3ca25cd854d16167fa7a67c91d24dadb8619c4543e: Status 404 returned error can't find the container with id c87f9d9af6ddab15e67e3d3ca25cd854d16167fa7a67c91d24dadb8619c4543e Nov 24 19:27:58 crc kubenswrapper[5035]: I1124 19:27:58.730225 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" event={"ID":"b782ab87-baed-4a09-b133-affcf2efd744","Type":"ContainerStarted","Data":"b26cfdf80c405eacc941bed3453dea6a10b4f776505d1ff2b679f377ac15d3b0"} Nov 24 19:27:58 crc kubenswrapper[5035]: I1124 19:27:58.731024 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" event={"ID":"60ba6695-0ae0-4dce-b6de-ec96a4723d07","Type":"ContainerStarted","Data":"c87f9d9af6ddab15e67e3d3ca25cd854d16167fa7a67c91d24dadb8619c4543e"} Nov 24 19:28:02 crc kubenswrapper[5035]: I1124 19:28:02.943625 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:28:02 crc kubenswrapper[5035]: I1124 19:28:02.944119 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:28:02 crc kubenswrapper[5035]: I1124 19:28:02.983579 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:28:03 crc kubenswrapper[5035]: I1124 19:28:03.855761 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:28:05 crc kubenswrapper[5035]: I1124 19:28:05.808653 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" event={"ID":"60ba6695-0ae0-4dce-b6de-ec96a4723d07","Type":"ContainerStarted","Data":"b1f857b2b7a3d7b519b0f0d8b175aaff543ab57eceafaa35a42ee47fc4b8701a"} Nov 24 19:28:05 crc kubenswrapper[5035]: I1124 19:28:05.808987 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:28:05 crc kubenswrapper[5035]: I1124 19:28:05.811154 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" event={"ID":"b782ab87-baed-4a09-b133-affcf2efd744","Type":"ContainerStarted","Data":"23f9578d1182275f29479a9149fd3f0297e972e9af46b0a6d4350584368a2877"} Nov 24 19:28:05 crc kubenswrapper[5035]: I1124 19:28:05.811210 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:28:05 crc kubenswrapper[5035]: I1124 19:28:05.833702 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" podStartSLOduration=2.202874983 podStartE2EDuration="8.833684572s" podCreationTimestamp="2025-11-24 19:27:57 +0000 UTC" firstStartedPulling="2025-11-24 19:27:58.405184786 +0000 UTC m=+816.927691043" lastFinishedPulling="2025-11-24 19:28:05.035994375 +0000 UTC m=+823.558500632" observedRunningTime="2025-11-24 19:28:05.829479107 +0000 UTC m=+824.351985384" watchObservedRunningTime="2025-11-24 19:28:05.833684572 +0000 UTC m=+824.356190849" Nov 24 19:28:05 crc kubenswrapper[5035]: I1124 19:28:05.857270 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" podStartSLOduration=1.9760558320000001 podStartE2EDuration="8.857242538s" podCreationTimestamp="2025-11-24 19:27:57 +0000 UTC" firstStartedPulling="2025-11-24 19:27:58.125587037 +0000 UTC m=+816.648093294" lastFinishedPulling="2025-11-24 19:28:05.006773743 +0000 UTC m=+823.529280000" observedRunningTime="2025-11-24 19:28:05.851959594 +0000 UTC m=+824.374465851" watchObservedRunningTime="2025-11-24 19:28:05.857242538 +0000 UTC m=+824.379748795" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.217541 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8chpn"] Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.217798 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8chpn" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="registry-server" containerID="cri-o://ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3" gracePeriod=2 Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.638844 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.750417 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-utilities\") pod \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.750532 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hszv4\" (UniqueName: \"kubernetes.io/projected/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-kube-api-access-hszv4\") pod \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.750588 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-catalog-content\") pod \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\" (UID: \"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2\") " Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.751338 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-utilities" (OuterVolumeSpecName: "utilities") pod "14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" (UID: "14dbbc19-6ce7-40b9-96a8-9638b06f3ba2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.756897 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-kube-api-access-hszv4" (OuterVolumeSpecName: "kube-api-access-hszv4") pod "14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" (UID: "14dbbc19-6ce7-40b9-96a8-9638b06f3ba2"). InnerVolumeSpecName "kube-api-access-hszv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.776349 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" (UID: "14dbbc19-6ce7-40b9-96a8-9638b06f3ba2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.818727 5035 generic.go:334] "Generic (PLEG): container finished" podID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerID="ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3" exitCode=0 Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.818765 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8chpn" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.818797 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8chpn" event={"ID":"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2","Type":"ContainerDied","Data":"ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3"} Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.818829 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8chpn" event={"ID":"14dbbc19-6ce7-40b9-96a8-9638b06f3ba2","Type":"ContainerDied","Data":"0168cb1f3d8b4286fad32c67f03dede3f62dea9bc2d5869aa252b5dc66cabbb0"} Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.818846 5035 scope.go:117] "RemoveContainer" containerID="ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.837268 5035 scope.go:117] "RemoveContainer" containerID="1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.852960 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.853006 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hszv4\" (UniqueName: \"kubernetes.io/projected/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-kube-api-access-hszv4\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.853019 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.856308 5035 scope.go:117] "RemoveContainer" containerID="481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.869538 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8chpn"] Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.872949 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8chpn"] Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.874234 5035 scope.go:117] "RemoveContainer" containerID="ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3" Nov 24 19:28:06 crc kubenswrapper[5035]: E1124 19:28:06.874568 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3\": container with ID starting with ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3 not found: ID does not exist" containerID="ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.874608 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3"} err="failed to get container status \"ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3\": rpc error: code = NotFound desc = could not find container \"ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3\": container with ID starting with ca76c921731b0ce1e18414d0719261b3b3f34b94c2182e58a3fb44bdd8d2c1b3 not found: ID does not exist" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.874631 5035 scope.go:117] "RemoveContainer" containerID="1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3" Nov 24 19:28:06 crc kubenswrapper[5035]: E1124 19:28:06.874893 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3\": container with ID starting with 1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3 not found: ID does not exist" containerID="1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.874929 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3"} err="failed to get container status \"1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3\": rpc error: code = NotFound desc = could not find container \"1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3\": container with ID starting with 1b1b13172d16a0517be38ae8394fb6bd0478a8420f8950293af91d6c8e4dcda3 not found: ID does not exist" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.874946 5035 scope.go:117] "RemoveContainer" containerID="481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e" Nov 24 19:28:06 crc kubenswrapper[5035]: E1124 19:28:06.875203 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e\": container with ID starting with 481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e not found: ID does not exist" containerID="481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e" Nov 24 19:28:06 crc kubenswrapper[5035]: I1124 19:28:06.875242 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e"} err="failed to get container status \"481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e\": rpc error: code = NotFound desc = could not find container \"481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e\": container with ID starting with 481c6363e224a500109241d5c2edfbfbca5a5d2ef69c9e1aeb5579028f98691e not found: ID does not exist" Nov 24 19:28:08 crc kubenswrapper[5035]: I1124 19:28:08.206101 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" path="/var/lib/kubelet/pods/14dbbc19-6ce7-40b9-96a8-9638b06f3ba2/volumes" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.230473 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-546bw"] Nov 24 19:28:16 crc kubenswrapper[5035]: E1124 19:28:16.231212 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="extract-utilities" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.231227 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="extract-utilities" Nov 24 19:28:16 crc kubenswrapper[5035]: E1124 19:28:16.231245 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="registry-server" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.231253 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="registry-server" Nov 24 19:28:16 crc kubenswrapper[5035]: E1124 19:28:16.231277 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="extract-content" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.231302 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="extract-content" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.231425 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="14dbbc19-6ce7-40b9-96a8-9638b06f3ba2" containerName="registry-server" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.232342 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.249308 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-546bw"] Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.374576 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4jrk\" (UniqueName: \"kubernetes.io/projected/f04c2d81-18ce-403c-82a4-a59495f96bd5-kube-api-access-t4jrk\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.374648 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-catalog-content\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.374681 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-utilities\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.476066 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-catalog-content\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.476142 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-utilities\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.476200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4jrk\" (UniqueName: \"kubernetes.io/projected/f04c2d81-18ce-403c-82a4-a59495f96bd5-kube-api-access-t4jrk\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.476625 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-catalog-content\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.476640 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-utilities\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.504274 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4jrk\" (UniqueName: \"kubernetes.io/projected/f04c2d81-18ce-403c-82a4-a59495f96bd5-kube-api-access-t4jrk\") pod \"community-operators-546bw\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.549064 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.849642 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-546bw"] Nov 24 19:28:16 crc kubenswrapper[5035]: I1124 19:28:16.885534 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-546bw" event={"ID":"f04c2d81-18ce-403c-82a4-a59495f96bd5","Type":"ContainerStarted","Data":"2a3efab826424bf9a772794a39c8016ca11a7933895d568f4509da3149d0e2ac"} Nov 24 19:28:17 crc kubenswrapper[5035]: I1124 19:28:17.918848 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6fbc66c766-4477t" Nov 24 19:28:18 crc kubenswrapper[5035]: I1124 19:28:18.901997 5035 generic.go:334] "Generic (PLEG): container finished" podID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerID="32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24" exitCode=0 Nov 24 19:28:18 crc kubenswrapper[5035]: I1124 19:28:18.902090 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-546bw" event={"ID":"f04c2d81-18ce-403c-82a4-a59495f96bd5","Type":"ContainerDied","Data":"32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24"} Nov 24 19:28:19 crc kubenswrapper[5035]: I1124 19:28:19.909175 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-546bw" event={"ID":"f04c2d81-18ce-403c-82a4-a59495f96bd5","Type":"ContainerStarted","Data":"145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f"} Nov 24 19:28:20 crc kubenswrapper[5035]: I1124 19:28:20.916433 5035 generic.go:334] "Generic (PLEG): container finished" podID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerID="145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f" exitCode=0 Nov 24 19:28:20 crc kubenswrapper[5035]: I1124 19:28:20.916507 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-546bw" event={"ID":"f04c2d81-18ce-403c-82a4-a59495f96bd5","Type":"ContainerDied","Data":"145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f"} Nov 24 19:28:21 crc kubenswrapper[5035]: I1124 19:28:21.927156 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-546bw" event={"ID":"f04c2d81-18ce-403c-82a4-a59495f96bd5","Type":"ContainerStarted","Data":"e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0"} Nov 24 19:28:25 crc kubenswrapper[5035]: I1124 19:28:25.741562 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-546bw" podStartSLOduration=7.064417899 podStartE2EDuration="9.741535923s" podCreationTimestamp="2025-11-24 19:28:16 +0000 UTC" firstStartedPulling="2025-11-24 19:28:18.9042085 +0000 UTC m=+837.426714757" lastFinishedPulling="2025-11-24 19:28:21.581326514 +0000 UTC m=+840.103832781" observedRunningTime="2025-11-24 19:28:21.950995903 +0000 UTC m=+840.473502170" watchObservedRunningTime="2025-11-24 19:28:25.741535923 +0000 UTC m=+844.264042220" Nov 24 19:28:25 crc kubenswrapper[5035]: I1124 19:28:25.744129 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-trs4x"] Nov 24 19:28:25 crc kubenswrapper[5035]: I1124 19:28:25.746604 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:25 crc kubenswrapper[5035]: I1124 19:28:25.756454 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-trs4x"] Nov 24 19:28:25 crc kubenswrapper[5035]: I1124 19:28:25.918483 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slcmj\" (UniqueName: \"kubernetes.io/projected/6ccb4b04-4593-4b23-b612-70a7ec78e763-kube-api-access-slcmj\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:25 crc kubenswrapper[5035]: I1124 19:28:25.919117 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-utilities\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:25 crc kubenswrapper[5035]: I1124 19:28:25.919246 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-catalog-content\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.020940 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-utilities\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.021009 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-catalog-content\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.021068 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slcmj\" (UniqueName: \"kubernetes.io/projected/6ccb4b04-4593-4b23-b612-70a7ec78e763-kube-api-access-slcmj\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.021753 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-utilities\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.021762 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-catalog-content\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.060367 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slcmj\" (UniqueName: \"kubernetes.io/projected/6ccb4b04-4593-4b23-b612-70a7ec78e763-kube-api-access-slcmj\") pod \"certified-operators-trs4x\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.079188 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.508426 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-trs4x"] Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.549265 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.549310 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.597585 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.954896 5035 generic.go:334] "Generic (PLEG): container finished" podID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerID="33300d992eb8b2b40df975a3571f8b41ae52156479562819f87c25e9bb1c3251" exitCode=0 Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.954969 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trs4x" event={"ID":"6ccb4b04-4593-4b23-b612-70a7ec78e763","Type":"ContainerDied","Data":"33300d992eb8b2b40df975a3571f8b41ae52156479562819f87c25e9bb1c3251"} Nov 24 19:28:26 crc kubenswrapper[5035]: I1124 19:28:26.955042 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trs4x" event={"ID":"6ccb4b04-4593-4b23-b612-70a7ec78e763","Type":"ContainerStarted","Data":"47953f257695e3f5ca0b97eb4d451d3d021260d3039faaf53e3648ed3f023a76"} Nov 24 19:28:27 crc kubenswrapper[5035]: I1124 19:28:27.032086 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:27 crc kubenswrapper[5035]: I1124 19:28:27.964170 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trs4x" event={"ID":"6ccb4b04-4593-4b23-b612-70a7ec78e763","Type":"ContainerStarted","Data":"8792d793c536e1ba62422e1010101073bc32c603810b0def8a08337bbe5bfc86"} Nov 24 19:28:28 crc kubenswrapper[5035]: I1124 19:28:28.978260 5035 generic.go:334] "Generic (PLEG): container finished" podID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerID="8792d793c536e1ba62422e1010101073bc32c603810b0def8a08337bbe5bfc86" exitCode=0 Nov 24 19:28:28 crc kubenswrapper[5035]: I1124 19:28:28.978392 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trs4x" event={"ID":"6ccb4b04-4593-4b23-b612-70a7ec78e763","Type":"ContainerDied","Data":"8792d793c536e1ba62422e1010101073bc32c603810b0def8a08337bbe5bfc86"} Nov 24 19:28:29 crc kubenswrapper[5035]: I1124 19:28:29.986451 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trs4x" event={"ID":"6ccb4b04-4593-4b23-b612-70a7ec78e763","Type":"ContainerStarted","Data":"c90411398eac7162b1269f97c74ba8197f471a86223f1a32adc933aa8f63e554"} Nov 24 19:28:30 crc kubenswrapper[5035]: I1124 19:28:30.007096 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-trs4x" podStartSLOduration=2.599121199 podStartE2EDuration="5.00706859s" podCreationTimestamp="2025-11-24 19:28:25 +0000 UTC" firstStartedPulling="2025-11-24 19:28:26.959661412 +0000 UTC m=+845.482167669" lastFinishedPulling="2025-11-24 19:28:29.367608783 +0000 UTC m=+847.890115060" observedRunningTime="2025-11-24 19:28:30.003485073 +0000 UTC m=+848.525991340" watchObservedRunningTime="2025-11-24 19:28:30.00706859 +0000 UTC m=+848.529574847" Nov 24 19:28:30 crc kubenswrapper[5035]: I1124 19:28:30.206887 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-546bw"] Nov 24 19:28:30 crc kubenswrapper[5035]: I1124 19:28:30.207406 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-546bw" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="registry-server" containerID="cri-o://e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0" gracePeriod=2 Nov 24 19:28:31 crc kubenswrapper[5035]: I1124 19:28:31.817225 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001023 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-utilities\") pod \"f04c2d81-18ce-403c-82a4-a59495f96bd5\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001111 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4jrk\" (UniqueName: \"kubernetes.io/projected/f04c2d81-18ce-403c-82a4-a59495f96bd5-kube-api-access-t4jrk\") pod \"f04c2d81-18ce-403c-82a4-a59495f96bd5\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001148 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-catalog-content\") pod \"f04c2d81-18ce-403c-82a4-a59495f96bd5\" (UID: \"f04c2d81-18ce-403c-82a4-a59495f96bd5\") " Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001168 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-546bw" event={"ID":"f04c2d81-18ce-403c-82a4-a59495f96bd5","Type":"ContainerDied","Data":"e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0"} Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001209 5035 scope.go:117] "RemoveContainer" containerID="e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001253 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-546bw" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001142 5035 generic.go:334] "Generic (PLEG): container finished" podID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerID="e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0" exitCode=0 Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.001373 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-546bw" event={"ID":"f04c2d81-18ce-403c-82a4-a59495f96bd5","Type":"ContainerDied","Data":"2a3efab826424bf9a772794a39c8016ca11a7933895d568f4509da3149d0e2ac"} Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.002039 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-utilities" (OuterVolumeSpecName: "utilities") pod "f04c2d81-18ce-403c-82a4-a59495f96bd5" (UID: "f04c2d81-18ce-403c-82a4-a59495f96bd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.009953 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04c2d81-18ce-403c-82a4-a59495f96bd5-kube-api-access-t4jrk" (OuterVolumeSpecName: "kube-api-access-t4jrk") pod "f04c2d81-18ce-403c-82a4-a59495f96bd5" (UID: "f04c2d81-18ce-403c-82a4-a59495f96bd5"). InnerVolumeSpecName "kube-api-access-t4jrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.043311 5035 scope.go:117] "RemoveContainer" containerID="145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.059506 5035 scope.go:117] "RemoveContainer" containerID="32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.060944 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f04c2d81-18ce-403c-82a4-a59495f96bd5" (UID: "f04c2d81-18ce-403c-82a4-a59495f96bd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.077664 5035 scope.go:117] "RemoveContainer" containerID="e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0" Nov 24 19:28:32 crc kubenswrapper[5035]: E1124 19:28:32.079110 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0\": container with ID starting with e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0 not found: ID does not exist" containerID="e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.079143 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0"} err="failed to get container status \"e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0\": rpc error: code = NotFound desc = could not find container \"e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0\": container with ID starting with e97c7b09a3220a34bc8e192d4f660254bde059e5af6ff77c4471929e4eeec5c0 not found: ID does not exist" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.079164 5035 scope.go:117] "RemoveContainer" containerID="145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f" Nov 24 19:28:32 crc kubenswrapper[5035]: E1124 19:28:32.079625 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f\": container with ID starting with 145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f not found: ID does not exist" containerID="145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.079658 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f"} err="failed to get container status \"145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f\": rpc error: code = NotFound desc = could not find container \"145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f\": container with ID starting with 145da93c5903010863fb85d58c57e3ae8aa6b8e0aab8cd637072a6ced8826d6f not found: ID does not exist" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.079676 5035 scope.go:117] "RemoveContainer" containerID="32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24" Nov 24 19:28:32 crc kubenswrapper[5035]: E1124 19:28:32.079992 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24\": container with ID starting with 32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24 not found: ID does not exist" containerID="32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.080039 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24"} err="failed to get container status \"32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24\": rpc error: code = NotFound desc = could not find container \"32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24\": container with ID starting with 32d32756ed0533e3b78f55f83f90ef8f15dd881b00e95070441299c25b70ee24 not found: ID does not exist" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.102900 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.102931 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4jrk\" (UniqueName: \"kubernetes.io/projected/f04c2d81-18ce-403c-82a4-a59495f96bd5-kube-api-access-t4jrk\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.102945 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04c2d81-18ce-403c-82a4-a59495f96bd5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.317700 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-546bw"] Nov 24 19:28:32 crc kubenswrapper[5035]: I1124 19:28:32.321814 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-546bw"] Nov 24 19:28:34 crc kubenswrapper[5035]: I1124 19:28:34.206519 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" path="/var/lib/kubelet/pods/f04c2d81-18ce-403c-82a4-a59495f96bd5/volumes" Nov 24 19:28:36 crc kubenswrapper[5035]: I1124 19:28:36.080028 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:36 crc kubenswrapper[5035]: I1124 19:28:36.080088 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:36 crc kubenswrapper[5035]: I1124 19:28:36.130815 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:37 crc kubenswrapper[5035]: I1124 19:28:37.079851 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:37 crc kubenswrapper[5035]: I1124 19:28:37.131009 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-trs4x"] Nov 24 19:28:37 crc kubenswrapper[5035]: I1124 19:28:37.632656 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6648cb8994-hqcpk" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.320055 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jx6sg"] Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.320388 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="extract-content" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.320403 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="extract-content" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.320417 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="extract-utilities" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.320424 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="extract-utilities" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.320433 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="registry-server" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.320440 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="registry-server" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.320582 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04c2d81-18ce-403c-82a4-a59495f96bd5" containerName="registry-server" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.322740 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.324373 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.324462 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-xbvbb" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.325100 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9drwf"] Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.325828 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.326056 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.327183 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.348201 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9drwf"] Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.385790 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics-certs\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.385848 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99zxb\" (UniqueName: \"kubernetes.io/projected/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-kube-api-access-99zxb\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.386012 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-sockets\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.386096 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.386154 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-reloader\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.386187 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4041fa2f-450a-4f60-baa7-15a8478934f4-cert\") pod \"frr-k8s-webhook-server-6998585d5-9drwf\" (UID: \"4041fa2f-450a-4f60-baa7-15a8478934f4\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.386277 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-startup\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.386350 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlrdg\" (UniqueName: \"kubernetes.io/projected/4041fa2f-450a-4f60-baa7-15a8478934f4-kube-api-access-dlrdg\") pod \"frr-k8s-webhook-server-6998585d5-9drwf\" (UID: \"4041fa2f-450a-4f60-baa7-15a8478934f4\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.386383 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-conf\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.427580 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-xxqk2"] Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.428758 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.430599 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.430864 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-lzp7n" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.432125 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.432981 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.443799 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-xpd9f"] Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.444723 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.446244 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.455129 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-xpd9f"] Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487490 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-reloader\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487561 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metallb-excludel2\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487583 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4041fa2f-450a-4f60-baa7-15a8478934f4-cert\") pod \"frr-k8s-webhook-server-6998585d5-9drwf\" (UID: \"4041fa2f-450a-4f60-baa7-15a8478934f4\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487605 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-cert\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487629 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-metrics-certs\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487651 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-startup\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487679 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8gx2\" (UniqueName: \"kubernetes.io/projected/785073c0-865d-4f23-814a-a393cd4caced-kube-api-access-f8gx2\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487699 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlrdg\" (UniqueName: \"kubernetes.io/projected/4041fa2f-450a-4f60-baa7-15a8478934f4-kube-api-access-dlrdg\") pod \"frr-k8s-webhook-server-6998585d5-9drwf\" (UID: \"4041fa2f-450a-4f60-baa7-15a8478934f4\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487718 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-conf\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.487734 5035 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.487812 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4041fa2f-450a-4f60-baa7-15a8478934f4-cert podName:4041fa2f-450a-4f60-baa7-15a8478934f4 nodeName:}" failed. No retries permitted until 2025-11-24 19:28:38.987790625 +0000 UTC m=+857.510296962 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4041fa2f-450a-4f60-baa7-15a8478934f4-cert") pod "frr-k8s-webhook-server-6998585d5-9drwf" (UID: "4041fa2f-450a-4f60-baa7-15a8478934f4") : secret "frr-k8s-webhook-server-cert" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487751 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb6mw\" (UniqueName: \"kubernetes.io/projected/4ae03b6d-7158-40a9-9b24-4a247860e4e1-kube-api-access-bb6mw\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487920 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics-certs\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487976 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99zxb\" (UniqueName: \"kubernetes.io/projected/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-kube-api-access-99zxb\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.487993 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-reloader\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488033 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.488046 5035 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488067 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-sockets\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488085 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-conf\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.488097 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics-certs podName:9003e7e1-21c5-41e2-8eff-a55dbeabaaa0 nodeName:}" failed. No retries permitted until 2025-11-24 19:28:38.988079514 +0000 UTC m=+857.510585771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics-certs") pod "frr-k8s-jx6sg" (UID: "9003e7e1-21c5-41e2-8eff-a55dbeabaaa0") : secret "frr-k8s-certs-secret" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488114 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metrics-certs\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488139 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488422 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-sockets\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488524 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.488755 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-frr-startup\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.515108 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99zxb\" (UniqueName: \"kubernetes.io/projected/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-kube-api-access-99zxb\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.515268 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlrdg\" (UniqueName: \"kubernetes.io/projected/4041fa2f-450a-4f60-baa7-15a8478934f4-kube-api-access-dlrdg\") pod \"frr-k8s-webhook-server-6998585d5-9drwf\" (UID: \"4041fa2f-450a-4f60-baa7-15a8478934f4\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.588659 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.588723 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metrics-certs\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.588763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metallb-excludel2\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.588768 5035 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.588803 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-cert\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.588830 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist podName:4ae03b6d-7158-40a9-9b24-4a247860e4e1 nodeName:}" failed. No retries permitted until 2025-11-24 19:28:39.088812697 +0000 UTC m=+857.611318954 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist") pod "speaker-xxqk2" (UID: "4ae03b6d-7158-40a9-9b24-4a247860e4e1") : secret "metallb-memberlist" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.588874 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-metrics-certs\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.588907 5035 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.588960 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8gx2\" (UniqueName: \"kubernetes.io/projected/785073c0-865d-4f23-814a-a393cd4caced-kube-api-access-f8gx2\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.588977 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metrics-certs podName:4ae03b6d-7158-40a9-9b24-4a247860e4e1 nodeName:}" failed. No retries permitted until 2025-11-24 19:28:39.088957221 +0000 UTC m=+857.611463588 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metrics-certs") pod "speaker-xxqk2" (UID: "4ae03b6d-7158-40a9-9b24-4a247860e4e1") : secret "speaker-certs-secret" not found Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.589031 5035 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.589039 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb6mw\" (UniqueName: \"kubernetes.io/projected/4ae03b6d-7158-40a9-9b24-4a247860e4e1-kube-api-access-bb6mw\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: E1124 19:28:38.589057 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-metrics-certs podName:785073c0-865d-4f23-814a-a393cd4caced nodeName:}" failed. No retries permitted until 2025-11-24 19:28:39.089049573 +0000 UTC m=+857.611555970 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-metrics-certs") pod "controller-6c7b4b5f48-xpd9f" (UID: "785073c0-865d-4f23-814a-a393cd4caced") : secret "controller-certs-secret" not found Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.589647 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metallb-excludel2\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.590494 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.603917 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-cert\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.605800 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb6mw\" (UniqueName: \"kubernetes.io/projected/4ae03b6d-7158-40a9-9b24-4a247860e4e1-kube-api-access-bb6mw\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.605811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8gx2\" (UniqueName: \"kubernetes.io/projected/785073c0-865d-4f23-814a-a393cd4caced-kube-api-access-f8gx2\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.993940 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics-certs\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.994943 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4041fa2f-450a-4f60-baa7-15a8478934f4-cert\") pod \"frr-k8s-webhook-server-6998585d5-9drwf\" (UID: \"4041fa2f-450a-4f60-baa7-15a8478934f4\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.997662 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9003e7e1-21c5-41e2-8eff-a55dbeabaaa0-metrics-certs\") pod \"frr-k8s-jx6sg\" (UID: \"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0\") " pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:38 crc kubenswrapper[5035]: I1124 19:28:38.998462 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4041fa2f-450a-4f60-baa7-15a8478934f4-cert\") pod \"frr-k8s-webhook-server-6998585d5-9drwf\" (UID: \"4041fa2f-450a-4f60-baa7-15a8478934f4\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.041385 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-trs4x" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="registry-server" containerID="cri-o://c90411398eac7162b1269f97c74ba8197f471a86223f1a32adc933aa8f63e554" gracePeriod=2 Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.096596 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.096646 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metrics-certs\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.096688 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-metrics-certs\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:39 crc kubenswrapper[5035]: E1124 19:28:39.096725 5035 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 19:28:39 crc kubenswrapper[5035]: E1124 19:28:39.096822 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist podName:4ae03b6d-7158-40a9-9b24-4a247860e4e1 nodeName:}" failed. No retries permitted until 2025-11-24 19:28:40.096801378 +0000 UTC m=+858.619307635 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist") pod "speaker-xxqk2" (UID: "4ae03b6d-7158-40a9-9b24-4a247860e4e1") : secret "metallb-memberlist" not found Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.100808 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-metrics-certs\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.101433 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/785073c0-865d-4f23-814a-a393cd4caced-metrics-certs\") pod \"controller-6c7b4b5f48-xpd9f\" (UID: \"785073c0-865d-4f23-814a-a393cd4caced\") " pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.242894 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.254871 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.357641 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.576049 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-xpd9f"] Nov 24 19:28:39 crc kubenswrapper[5035]: I1124 19:28:39.692131 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-9drwf"] Nov 24 19:28:39 crc kubenswrapper[5035]: W1124 19:28:39.702456 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4041fa2f_450a_4f60_baa7_15a8478934f4.slice/crio-889feefa72687ce5685cc8c5fabedcd962348f93f3131588a66b7df59806deef WatchSource:0}: Error finding container 889feefa72687ce5685cc8c5fabedcd962348f93f3131588a66b7df59806deef: Status 404 returned error can't find the container with id 889feefa72687ce5685cc8c5fabedcd962348f93f3131588a66b7df59806deef Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.048991 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" event={"ID":"4041fa2f-450a-4f60-baa7-15a8478934f4","Type":"ContainerStarted","Data":"889feefa72687ce5685cc8c5fabedcd962348f93f3131588a66b7df59806deef"} Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.049871 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerStarted","Data":"0502bac3e1062cbb6b6fdbf7570d663c67df39d5cce93370b3e70047716f23f1"} Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.052327 5035 generic.go:334] "Generic (PLEG): container finished" podID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerID="c90411398eac7162b1269f97c74ba8197f471a86223f1a32adc933aa8f63e554" exitCode=0 Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.052380 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trs4x" event={"ID":"6ccb4b04-4593-4b23-b612-70a7ec78e763","Type":"ContainerDied","Data":"c90411398eac7162b1269f97c74ba8197f471a86223f1a32adc933aa8f63e554"} Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.053894 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-xpd9f" event={"ID":"785073c0-865d-4f23-814a-a393cd4caced","Type":"ContainerStarted","Data":"73303aa1ada5ad2fdbe3a94807899b211dc12516bf8b2ea22640903bf21cf577"} Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.053917 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-xpd9f" event={"ID":"785073c0-865d-4f23-814a-a393cd4caced","Type":"ContainerStarted","Data":"cc55ddb572a3d1a3e15b36d8dc150abd3470e1ab204b823768f10b431283b06b"} Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.107124 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.112100 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/4ae03b6d-7158-40a9-9b24-4a247860e4e1-memberlist\") pod \"speaker-xxqk2\" (UID: \"4ae03b6d-7158-40a9-9b24-4a247860e4e1\") " pod="metallb-system/speaker-xxqk2" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.241760 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xxqk2" Nov 24 19:28:40 crc kubenswrapper[5035]: W1124 19:28:40.256484 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ae03b6d_7158_40a9_9b24_4a247860e4e1.slice/crio-ebff1c6ff5f4ddb34ce8d063bb5fda182c0dd4a6ccac0beb4490895aa54f63cb WatchSource:0}: Error finding container ebff1c6ff5f4ddb34ce8d063bb5fda182c0dd4a6ccac0beb4490895aa54f63cb: Status 404 returned error can't find the container with id ebff1c6ff5f4ddb34ce8d063bb5fda182c0dd4a6ccac0beb4490895aa54f63cb Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.587794 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.722902 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slcmj\" (UniqueName: \"kubernetes.io/projected/6ccb4b04-4593-4b23-b612-70a7ec78e763-kube-api-access-slcmj\") pod \"6ccb4b04-4593-4b23-b612-70a7ec78e763\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.722969 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-catalog-content\") pod \"6ccb4b04-4593-4b23-b612-70a7ec78e763\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.723012 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-utilities\") pod \"6ccb4b04-4593-4b23-b612-70a7ec78e763\" (UID: \"6ccb4b04-4593-4b23-b612-70a7ec78e763\") " Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.724035 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-utilities" (OuterVolumeSpecName: "utilities") pod "6ccb4b04-4593-4b23-b612-70a7ec78e763" (UID: "6ccb4b04-4593-4b23-b612-70a7ec78e763"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.736755 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ccb4b04-4593-4b23-b612-70a7ec78e763-kube-api-access-slcmj" (OuterVolumeSpecName: "kube-api-access-slcmj") pod "6ccb4b04-4593-4b23-b612-70a7ec78e763" (UID: "6ccb4b04-4593-4b23-b612-70a7ec78e763"). InnerVolumeSpecName "kube-api-access-slcmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.769502 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ccb4b04-4593-4b23-b612-70a7ec78e763" (UID: "6ccb4b04-4593-4b23-b612-70a7ec78e763"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.825818 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.826183 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ccb4b04-4593-4b23-b612-70a7ec78e763-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:40 crc kubenswrapper[5035]: I1124 19:28:40.826201 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slcmj\" (UniqueName: \"kubernetes.io/projected/6ccb4b04-4593-4b23-b612-70a7ec78e763-kube-api-access-slcmj\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.073127 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-trs4x" event={"ID":"6ccb4b04-4593-4b23-b612-70a7ec78e763","Type":"ContainerDied","Data":"47953f257695e3f5ca0b97eb4d451d3d021260d3039faaf53e3648ed3f023a76"} Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.073176 5035 scope.go:117] "RemoveContainer" containerID="c90411398eac7162b1269f97c74ba8197f471a86223f1a32adc933aa8f63e554" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.073137 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-trs4x" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.079603 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xxqk2" event={"ID":"4ae03b6d-7158-40a9-9b24-4a247860e4e1","Type":"ContainerStarted","Data":"69c4516567bd80feb5e305a317aeac5498416eec342537274126a135200d74b7"} Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.079652 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xxqk2" event={"ID":"4ae03b6d-7158-40a9-9b24-4a247860e4e1","Type":"ContainerStarted","Data":"f6cd0c4e269447898f241aa99c7f5510d408726168d888022ec8126b05582389"} Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.079666 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xxqk2" event={"ID":"4ae03b6d-7158-40a9-9b24-4a247860e4e1","Type":"ContainerStarted","Data":"ebff1c6ff5f4ddb34ce8d063bb5fda182c0dd4a6ccac0beb4490895aa54f63cb"} Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.080231 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-xxqk2" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.102030 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-xpd9f" event={"ID":"785073c0-865d-4f23-814a-a393cd4caced","Type":"ContainerStarted","Data":"85743f1bc6c778b114e1cb6cb859cede5b9ab3664e06f6e3b43bdbd091ef8d85"} Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.102730 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.104630 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-xxqk2" podStartSLOduration=3.104605125 podStartE2EDuration="3.104605125s" podCreationTimestamp="2025-11-24 19:28:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:28:41.101763787 +0000 UTC m=+859.624270044" watchObservedRunningTime="2025-11-24 19:28:41.104605125 +0000 UTC m=+859.627111382" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.116676 5035 scope.go:117] "RemoveContainer" containerID="8792d793c536e1ba62422e1010101073bc32c603810b0def8a08337bbe5bfc86" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.117080 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-trs4x"] Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.122321 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-trs4x"] Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.140252 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-xpd9f" podStartSLOduration=3.140237413 podStartE2EDuration="3.140237413s" podCreationTimestamp="2025-11-24 19:28:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:28:41.138483345 +0000 UTC m=+859.660989602" watchObservedRunningTime="2025-11-24 19:28:41.140237413 +0000 UTC m=+859.662743670" Nov 24 19:28:41 crc kubenswrapper[5035]: I1124 19:28:41.149937 5035 scope.go:117] "RemoveContainer" containerID="33300d992eb8b2b40df975a3571f8b41ae52156479562819f87c25e9bb1c3251" Nov 24 19:28:42 crc kubenswrapper[5035]: I1124 19:28:42.213172 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" path="/var/lib/kubelet/pods/6ccb4b04-4593-4b23-b612-70a7ec78e763/volumes" Nov 24 19:28:47 crc kubenswrapper[5035]: I1124 19:28:47.157545 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" event={"ID":"4041fa2f-450a-4f60-baa7-15a8478934f4","Type":"ContainerStarted","Data":"53a8a4281f99601525cffcbde71c24bc962684907ce72a23b6837515d4d90b22"} Nov 24 19:28:47 crc kubenswrapper[5035]: I1124 19:28:47.159761 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:47 crc kubenswrapper[5035]: I1124 19:28:47.160871 5035 generic.go:334] "Generic (PLEG): container finished" podID="9003e7e1-21c5-41e2-8eff-a55dbeabaaa0" containerID="154ed1fe577a39c059f50e667c681f9b746c64a6ed293e2edc50244101b93f8e" exitCode=0 Nov 24 19:28:47 crc kubenswrapper[5035]: I1124 19:28:47.160922 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerDied","Data":"154ed1fe577a39c059f50e667c681f9b746c64a6ed293e2edc50244101b93f8e"} Nov 24 19:28:47 crc kubenswrapper[5035]: I1124 19:28:47.199779 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" podStartSLOduration=2.6774623 podStartE2EDuration="9.199752984s" podCreationTimestamp="2025-11-24 19:28:38 +0000 UTC" firstStartedPulling="2025-11-24 19:28:39.704202777 +0000 UTC m=+858.226709034" lastFinishedPulling="2025-11-24 19:28:46.226493421 +0000 UTC m=+864.748999718" observedRunningTime="2025-11-24 19:28:47.181686478 +0000 UTC m=+865.704192775" watchObservedRunningTime="2025-11-24 19:28:47.199752984 +0000 UTC m=+865.722259251" Nov 24 19:28:48 crc kubenswrapper[5035]: I1124 19:28:48.168798 5035 generic.go:334] "Generic (PLEG): container finished" podID="9003e7e1-21c5-41e2-8eff-a55dbeabaaa0" containerID="863692d0063c7b9415a17c100709d8d94290e735a94c585b708ea0a04bd71174" exitCode=0 Nov 24 19:28:48 crc kubenswrapper[5035]: I1124 19:28:48.168912 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerDied","Data":"863692d0063c7b9415a17c100709d8d94290e735a94c585b708ea0a04bd71174"} Nov 24 19:28:49 crc kubenswrapper[5035]: I1124 19:28:49.180475 5035 generic.go:334] "Generic (PLEG): container finished" podID="9003e7e1-21c5-41e2-8eff-a55dbeabaaa0" containerID="c13243d42eaf246726863b62038a1b6a2672b15257def99df053ba10c58c1ac9" exitCode=0 Nov 24 19:28:49 crc kubenswrapper[5035]: I1124 19:28:49.180584 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerDied","Data":"c13243d42eaf246726863b62038a1b6a2672b15257def99df053ba10c58c1ac9"} Nov 24 19:28:49 crc kubenswrapper[5035]: I1124 19:28:49.361800 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-xpd9f" Nov 24 19:28:50 crc kubenswrapper[5035]: I1124 19:28:50.198187 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerStarted","Data":"352d74d18bd87aa99a95fa28bfa24d155915bc8f7ebed49158e5bd31f73ca2a3"} Nov 24 19:28:50 crc kubenswrapper[5035]: I1124 19:28:50.198515 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerStarted","Data":"903d50e29316c75ce9fab796177ec94f015b05cf628ed6fb1034cae9ad489b2c"} Nov 24 19:28:50 crc kubenswrapper[5035]: I1124 19:28:50.198530 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerStarted","Data":"b5be45d34c05922896d93dd581a4d6e5005c79797a5213832b141dacce14e405"} Nov 24 19:28:50 crc kubenswrapper[5035]: I1124 19:28:50.198541 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerStarted","Data":"96aedcf584a2ff05a198c8a6109905f0e3965b54e84af87abca8be6c08bba392"} Nov 24 19:28:50 crc kubenswrapper[5035]: I1124 19:28:50.198554 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerStarted","Data":"569e41ced124ff2dfbc106cf76b2b4886cf942426c8b86173b55c8f53d05a18a"} Nov 24 19:28:50 crc kubenswrapper[5035]: I1124 19:28:50.247351 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-xxqk2" Nov 24 19:28:51 crc kubenswrapper[5035]: I1124 19:28:51.209110 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jx6sg" event={"ID":"9003e7e1-21c5-41e2-8eff-a55dbeabaaa0","Type":"ContainerStarted","Data":"f608823ba59c84deac6a951163bfa13d19d95bb22da12b301c9078b87881f6d6"} Nov 24 19:28:51 crc kubenswrapper[5035]: I1124 19:28:51.209342 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:51 crc kubenswrapper[5035]: I1124 19:28:51.230671 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jx6sg" podStartSLOduration=6.546140806 podStartE2EDuration="13.230653527s" podCreationTimestamp="2025-11-24 19:28:38 +0000 UTC" firstStartedPulling="2025-11-24 19:28:39.570901262 +0000 UTC m=+858.093407559" lastFinishedPulling="2025-11-24 19:28:46.255414023 +0000 UTC m=+864.777920280" observedRunningTime="2025-11-24 19:28:51.228758694 +0000 UTC m=+869.751264961" watchObservedRunningTime="2025-11-24 19:28:51.230653527 +0000 UTC m=+869.753159784" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.106170 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mnqw8"] Nov 24 19:28:53 crc kubenswrapper[5035]: E1124 19:28:53.106666 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="registry-server" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.106678 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="registry-server" Nov 24 19:28:53 crc kubenswrapper[5035]: E1124 19:28:53.106699 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="extract-content" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.106704 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="extract-content" Nov 24 19:28:53 crc kubenswrapper[5035]: E1124 19:28:53.106714 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="extract-utilities" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.106720 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="extract-utilities" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.106813 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ccb4b04-4593-4b23-b612-70a7ec78e763" containerName="registry-server" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.107216 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mnqw8" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.109028 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-v8kzl" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.114278 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4nnl\" (UniqueName: \"kubernetes.io/projected/21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d-kube-api-access-z4nnl\") pod \"openstack-operator-index-mnqw8\" (UID: \"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d\") " pod="openstack-operators/openstack-operator-index-mnqw8" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.115411 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.118397 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.124452 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mnqw8"] Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.215188 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4nnl\" (UniqueName: \"kubernetes.io/projected/21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d-kube-api-access-z4nnl\") pod \"openstack-operator-index-mnqw8\" (UID: \"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d\") " pod="openstack-operators/openstack-operator-index-mnqw8" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.245991 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4nnl\" (UniqueName: \"kubernetes.io/projected/21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d-kube-api-access-z4nnl\") pod \"openstack-operator-index-mnqw8\" (UID: \"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d\") " pod="openstack-operators/openstack-operator-index-mnqw8" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.422007 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mnqw8" Nov 24 19:28:53 crc kubenswrapper[5035]: I1124 19:28:53.828452 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mnqw8"] Nov 24 19:28:53 crc kubenswrapper[5035]: W1124 19:28:53.847337 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21ff2f38_58c5_4ce2_a82c_f9c4002a9e6d.slice/crio-49def67fe12f3622dbd8db63f72b5007c403039f64f0ead8b6e187a7929bdfb2 WatchSource:0}: Error finding container 49def67fe12f3622dbd8db63f72b5007c403039f64f0ead8b6e187a7929bdfb2: Status 404 returned error can't find the container with id 49def67fe12f3622dbd8db63f72b5007c403039f64f0ead8b6e187a7929bdfb2 Nov 24 19:28:54 crc kubenswrapper[5035]: I1124 19:28:54.223727 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mnqw8" event={"ID":"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d","Type":"ContainerStarted","Data":"49def67fe12f3622dbd8db63f72b5007c403039f64f0ead8b6e187a7929bdfb2"} Nov 24 19:28:54 crc kubenswrapper[5035]: I1124 19:28:54.243000 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:54 crc kubenswrapper[5035]: I1124 19:28:54.288757 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:55 crc kubenswrapper[5035]: I1124 19:28:55.242372 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mnqw8" event={"ID":"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d","Type":"ContainerStarted","Data":"594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b"} Nov 24 19:28:55 crc kubenswrapper[5035]: I1124 19:28:55.261754 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mnqw8" podStartSLOduration=1.273758529 podStartE2EDuration="2.261734685s" podCreationTimestamp="2025-11-24 19:28:53 +0000 UTC" firstStartedPulling="2025-11-24 19:28:53.849355228 +0000 UTC m=+872.371861485" lastFinishedPulling="2025-11-24 19:28:54.837331364 +0000 UTC m=+873.359837641" observedRunningTime="2025-11-24 19:28:55.258543627 +0000 UTC m=+873.781049944" watchObservedRunningTime="2025-11-24 19:28:55.261734685 +0000 UTC m=+873.784240952" Nov 24 19:28:56 crc kubenswrapper[5035]: I1124 19:28:56.686042 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mnqw8"] Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.252852 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-mnqw8" podUID="21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d" containerName="registry-server" containerID="cri-o://594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b" gracePeriod=2 Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.488327 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-lkrfx"] Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.489821 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.508429 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lkrfx"] Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.670146 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x64xb\" (UniqueName: \"kubernetes.io/projected/25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9-kube-api-access-x64xb\") pod \"openstack-operator-index-lkrfx\" (UID: \"25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9\") " pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.670643 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mnqw8" Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.771909 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x64xb\" (UniqueName: \"kubernetes.io/projected/25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9-kube-api-access-x64xb\") pod \"openstack-operator-index-lkrfx\" (UID: \"25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9\") " pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.796658 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x64xb\" (UniqueName: \"kubernetes.io/projected/25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9-kube-api-access-x64xb\") pod \"openstack-operator-index-lkrfx\" (UID: \"25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9\") " pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.830254 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.873807 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4nnl\" (UniqueName: \"kubernetes.io/projected/21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d-kube-api-access-z4nnl\") pod \"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d\" (UID: \"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d\") " Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.878325 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d-kube-api-access-z4nnl" (OuterVolumeSpecName: "kube-api-access-z4nnl") pod "21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d" (UID: "21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d"). InnerVolumeSpecName "kube-api-access-z4nnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:28:57 crc kubenswrapper[5035]: I1124 19:28:57.975127 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4nnl\" (UniqueName: \"kubernetes.io/projected/21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d-kube-api-access-z4nnl\") on node \"crc\" DevicePath \"\"" Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.053226 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lkrfx"] Nov 24 19:28:58 crc kubenswrapper[5035]: W1124 19:28:58.056361 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25e32e21_6ff4_4b8f_aa65_23c9cff7b8c9.slice/crio-3d9160d9c59608984161b79beff570c68541e76100151895d09a3fa20f42d246 WatchSource:0}: Error finding container 3d9160d9c59608984161b79beff570c68541e76100151895d09a3fa20f42d246: Status 404 returned error can't find the container with id 3d9160d9c59608984161b79beff570c68541e76100151895d09a3fa20f42d246 Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.261386 5035 generic.go:334] "Generic (PLEG): container finished" podID="21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d" containerID="594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b" exitCode=0 Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.261458 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mnqw8" Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.261470 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mnqw8" event={"ID":"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d","Type":"ContainerDied","Data":"594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b"} Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.261529 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mnqw8" event={"ID":"21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d","Type":"ContainerDied","Data":"49def67fe12f3622dbd8db63f72b5007c403039f64f0ead8b6e187a7929bdfb2"} Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.261554 5035 scope.go:117] "RemoveContainer" containerID="594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b" Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.264353 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lkrfx" event={"ID":"25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9","Type":"ContainerStarted","Data":"3d9160d9c59608984161b79beff570c68541e76100151895d09a3fa20f42d246"} Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.299691 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mnqw8"] Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.303787 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-mnqw8"] Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.404562 5035 scope.go:117] "RemoveContainer" containerID="594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b" Nov 24 19:28:58 crc kubenswrapper[5035]: E1124 19:28:58.404996 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b\": container with ID starting with 594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b not found: ID does not exist" containerID="594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b" Nov 24 19:28:58 crc kubenswrapper[5035]: I1124 19:28:58.405034 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b"} err="failed to get container status \"594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b\": rpc error: code = NotFound desc = could not find container \"594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b\": container with ID starting with 594d08dd2fb976af5d48814e0c1c64b4b3ec4186296b1453cfe4b3a7f7ed358b not found: ID does not exist" Nov 24 19:28:59 crc kubenswrapper[5035]: I1124 19:28:59.248055 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jx6sg" Nov 24 19:28:59 crc kubenswrapper[5035]: I1124 19:28:59.264111 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-9drwf" Nov 24 19:28:59 crc kubenswrapper[5035]: I1124 19:28:59.271176 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lkrfx" event={"ID":"25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9","Type":"ContainerStarted","Data":"0efffcfa534ee277dfc4e71687f91d68e4f37f2f9e6a063fc79c844cc5b0d532"} Nov 24 19:28:59 crc kubenswrapper[5035]: I1124 19:28:59.292771 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-lkrfx" podStartSLOduration=1.475511528 podStartE2EDuration="2.292748161s" podCreationTimestamp="2025-11-24 19:28:57 +0000 UTC" firstStartedPulling="2025-11-24 19:28:58.060794123 +0000 UTC m=+876.583300380" lastFinishedPulling="2025-11-24 19:28:58.878030746 +0000 UTC m=+877.400537013" observedRunningTime="2025-11-24 19:28:59.290831688 +0000 UTC m=+877.813337965" watchObservedRunningTime="2025-11-24 19:28:59.292748161 +0000 UTC m=+877.815254418" Nov 24 19:29:00 crc kubenswrapper[5035]: I1124 19:29:00.208152 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d" path="/var/lib/kubelet/pods/21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d/volumes" Nov 24 19:29:07 crc kubenswrapper[5035]: I1124 19:29:07.830629 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:29:07 crc kubenswrapper[5035]: I1124 19:29:07.831057 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:29:07 crc kubenswrapper[5035]: I1124 19:29:07.861105 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:29:08 crc kubenswrapper[5035]: I1124 19:29:08.362325 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-lkrfx" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.332810 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp"] Nov 24 19:29:10 crc kubenswrapper[5035]: E1124 19:29:10.334169 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d" containerName="registry-server" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.334272 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d" containerName="registry-server" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.334722 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="21ff2f38-58c5-4ce2-a82c-f9c4002a9e6d" containerName="registry-server" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.335861 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.338700 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7t57w" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.389889 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.389946 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rrq4\" (UniqueName: \"kubernetes.io/projected/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-kube-api-access-4rrq4\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.390086 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.401856 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp"] Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.491511 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.491649 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rrq4\" (UniqueName: \"kubernetes.io/projected/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-kube-api-access-4rrq4\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.491873 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.492149 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.492660 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.518423 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rrq4\" (UniqueName: \"kubernetes.io/projected/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-kube-api-access-4rrq4\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:10 crc kubenswrapper[5035]: I1124 19:29:10.711331 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:11 crc kubenswrapper[5035]: I1124 19:29:11.165411 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp"] Nov 24 19:29:11 crc kubenswrapper[5035]: W1124 19:29:11.176400 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod089631e1_99a1_47f0_ac8f_d42bfb4f2b9f.slice/crio-9c5acdbc906d2cebffbc2b34a73d37f992dd865bb6870d4f45984621a4da0821 WatchSource:0}: Error finding container 9c5acdbc906d2cebffbc2b34a73d37f992dd865bb6870d4f45984621a4da0821: Status 404 returned error can't find the container with id 9c5acdbc906d2cebffbc2b34a73d37f992dd865bb6870d4f45984621a4da0821 Nov 24 19:29:11 crc kubenswrapper[5035]: I1124 19:29:11.403205 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" event={"ID":"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f","Type":"ContainerStarted","Data":"a9b93795f1e6f7701e441dd16d22322cd1f0ff4779e09fe0b444c89fef446a00"} Nov 24 19:29:11 crc kubenswrapper[5035]: I1124 19:29:11.403253 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" event={"ID":"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f","Type":"ContainerStarted","Data":"9c5acdbc906d2cebffbc2b34a73d37f992dd865bb6870d4f45984621a4da0821"} Nov 24 19:29:12 crc kubenswrapper[5035]: I1124 19:29:12.412559 5035 generic.go:334] "Generic (PLEG): container finished" podID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerID="a9b93795f1e6f7701e441dd16d22322cd1f0ff4779e09fe0b444c89fef446a00" exitCode=0 Nov 24 19:29:12 crc kubenswrapper[5035]: I1124 19:29:12.412615 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" event={"ID":"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f","Type":"ContainerDied","Data":"a9b93795f1e6f7701e441dd16d22322cd1f0ff4779e09fe0b444c89fef446a00"} Nov 24 19:29:14 crc kubenswrapper[5035]: I1124 19:29:14.433453 5035 generic.go:334] "Generic (PLEG): container finished" podID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerID="756fe66fb01ef4f5898ccb82669fc5044462701fa2fbf5e18f3e0934bb03fdda" exitCode=0 Nov 24 19:29:14 crc kubenswrapper[5035]: I1124 19:29:14.433529 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" event={"ID":"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f","Type":"ContainerDied","Data":"756fe66fb01ef4f5898ccb82669fc5044462701fa2fbf5e18f3e0934bb03fdda"} Nov 24 19:29:15 crc kubenswrapper[5035]: I1124 19:29:15.445761 5035 generic.go:334] "Generic (PLEG): container finished" podID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerID="12c8cee371f4114aeafc44bcb13f8cb539143b1fb14676fac1bbcaab42f6444d" exitCode=0 Nov 24 19:29:15 crc kubenswrapper[5035]: I1124 19:29:15.445830 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" event={"ID":"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f","Type":"ContainerDied","Data":"12c8cee371f4114aeafc44bcb13f8cb539143b1fb14676fac1bbcaab42f6444d"} Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.768613 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.879252 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rrq4\" (UniqueName: \"kubernetes.io/projected/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-kube-api-access-4rrq4\") pod \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.881411 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-util\") pod \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.881497 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-bundle\") pod \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\" (UID: \"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f\") " Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.882109 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-bundle" (OuterVolumeSpecName: "bundle") pod "089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" (UID: "089631e1-99a1-47f0-ac8f-d42bfb4f2b9f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.886733 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-kube-api-access-4rrq4" (OuterVolumeSpecName: "kube-api-access-4rrq4") pod "089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" (UID: "089631e1-99a1-47f0-ac8f-d42bfb4f2b9f"). InnerVolumeSpecName "kube-api-access-4rrq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.997180 5035 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:29:16 crc kubenswrapper[5035]: I1124 19:29:16.997256 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rrq4\" (UniqueName: \"kubernetes.io/projected/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-kube-api-access-4rrq4\") on node \"crc\" DevicePath \"\"" Nov 24 19:29:17 crc kubenswrapper[5035]: I1124 19:29:17.164854 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-util" (OuterVolumeSpecName: "util") pod "089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" (UID: "089631e1-99a1-47f0-ac8f-d42bfb4f2b9f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:29:17 crc kubenswrapper[5035]: I1124 19:29:17.200229 5035 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/089631e1-99a1-47f0-ac8f-d42bfb4f2b9f-util\") on node \"crc\" DevicePath \"\"" Nov 24 19:29:17 crc kubenswrapper[5035]: I1124 19:29:17.466638 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" event={"ID":"089631e1-99a1-47f0-ac8f-d42bfb4f2b9f","Type":"ContainerDied","Data":"9c5acdbc906d2cebffbc2b34a73d37f992dd865bb6870d4f45984621a4da0821"} Nov 24 19:29:17 crc kubenswrapper[5035]: I1124 19:29:17.466726 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c5acdbc906d2cebffbc2b34a73d37f992dd865bb6870d4f45984621a4da0821" Nov 24 19:29:17 crc kubenswrapper[5035]: I1124 19:29:17.466845 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.907085 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q"] Nov 24 19:29:22 crc kubenswrapper[5035]: E1124 19:29:22.907919 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerName="extract" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.907935 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerName="extract" Nov 24 19:29:22 crc kubenswrapper[5035]: E1124 19:29:22.907960 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerName="util" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.907971 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerName="util" Nov 24 19:29:22 crc kubenswrapper[5035]: E1124 19:29:22.907982 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerName="pull" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.907990 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerName="pull" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.908133 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="089631e1-99a1-47f0-ac8f-d42bfb4f2b9f" containerName="extract" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.908610 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.910370 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-wnwzc" Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.932058 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q"] Nov 24 19:29:22 crc kubenswrapper[5035]: I1124 19:29:22.978064 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zmvb\" (UniqueName: \"kubernetes.io/projected/e33191d3-0eda-4de9-80d6-cd5d077d1862-kube-api-access-5zmvb\") pod \"openstack-operator-controller-operator-7b567956b5-6tq5q\" (UID: \"e33191d3-0eda-4de9-80d6-cd5d077d1862\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 19:29:23 crc kubenswrapper[5035]: I1124 19:29:23.079323 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zmvb\" (UniqueName: \"kubernetes.io/projected/e33191d3-0eda-4de9-80d6-cd5d077d1862-kube-api-access-5zmvb\") pod \"openstack-operator-controller-operator-7b567956b5-6tq5q\" (UID: \"e33191d3-0eda-4de9-80d6-cd5d077d1862\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 19:29:23 crc kubenswrapper[5035]: I1124 19:29:23.096923 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zmvb\" (UniqueName: \"kubernetes.io/projected/e33191d3-0eda-4de9-80d6-cd5d077d1862-kube-api-access-5zmvb\") pod \"openstack-operator-controller-operator-7b567956b5-6tq5q\" (UID: \"e33191d3-0eda-4de9-80d6-cd5d077d1862\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 19:29:23 crc kubenswrapper[5035]: I1124 19:29:23.226878 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 19:29:23 crc kubenswrapper[5035]: I1124 19:29:23.466354 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q"] Nov 24 19:29:23 crc kubenswrapper[5035]: W1124 19:29:23.481626 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode33191d3_0eda_4de9_80d6_cd5d077d1862.slice/crio-3ad7983857682722e58494ce2552fc21c7c4b0231239892f1e36cccf7f552b2c WatchSource:0}: Error finding container 3ad7983857682722e58494ce2552fc21c7c4b0231239892f1e36cccf7f552b2c: Status 404 returned error can't find the container with id 3ad7983857682722e58494ce2552fc21c7c4b0231239892f1e36cccf7f552b2c Nov 24 19:29:23 crc kubenswrapper[5035]: I1124 19:29:23.505178 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" event={"ID":"e33191d3-0eda-4de9-80d6-cd5d077d1862","Type":"ContainerStarted","Data":"3ad7983857682722e58494ce2552fc21c7c4b0231239892f1e36cccf7f552b2c"} Nov 24 19:29:29 crc kubenswrapper[5035]: I1124 19:29:29.558559 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" event={"ID":"e33191d3-0eda-4de9-80d6-cd5d077d1862","Type":"ContainerStarted","Data":"3d257b51d15f63f9b8e4558df03bdc2cea33f02fba8ec4c3a9b862d47e01c619"} Nov 24 19:29:29 crc kubenswrapper[5035]: I1124 19:29:29.559276 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 19:29:29 crc kubenswrapper[5035]: I1124 19:29:29.603277 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" podStartSLOduration=2.253327139 podStartE2EDuration="7.603248768s" podCreationTimestamp="2025-11-24 19:29:22 +0000 UTC" firstStartedPulling="2025-11-24 19:29:23.494732393 +0000 UTC m=+902.017238650" lastFinishedPulling="2025-11-24 19:29:28.844654002 +0000 UTC m=+907.367160279" observedRunningTime="2025-11-24 19:29:29.60007562 +0000 UTC m=+908.122581887" watchObservedRunningTime="2025-11-24 19:29:29.603248768 +0000 UTC m=+908.125755065" Nov 24 19:29:43 crc kubenswrapper[5035]: I1124 19:29:43.228988 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 19:29:45 crc kubenswrapper[5035]: I1124 19:29:45.234182 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:29:45 crc kubenswrapper[5035]: I1124 19:29:45.234256 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.124623 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.125879 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.127883 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.128134 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.134828 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.240142 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e46e48-d877-4332-918a-95d8c6afc3b5-secret-volume\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.240200 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7lnl\" (UniqueName: \"kubernetes.io/projected/b3e46e48-d877-4332-918a-95d8c6afc3b5-kube-api-access-g7lnl\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.240443 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e46e48-d877-4332-918a-95d8c6afc3b5-config-volume\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.342186 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e46e48-d877-4332-918a-95d8c6afc3b5-secret-volume\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.342245 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7lnl\" (UniqueName: \"kubernetes.io/projected/b3e46e48-d877-4332-918a-95d8c6afc3b5-kube-api-access-g7lnl\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.342281 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e46e48-d877-4332-918a-95d8c6afc3b5-config-volume\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.343140 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e46e48-d877-4332-918a-95d8c6afc3b5-config-volume\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.354908 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e46e48-d877-4332-918a-95d8c6afc3b5-secret-volume\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.359753 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7lnl\" (UniqueName: \"kubernetes.io/projected/b3e46e48-d877-4332-918a-95d8c6afc3b5-kube-api-access-g7lnl\") pod \"collect-profiles-29400210-26p8w\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.440411 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.554346 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.557082 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.572159 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fbwfb" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.581463 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.609970 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.611188 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.616185 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.617467 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.621982 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-6xgt7" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.622804 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-chsvt" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.628650 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.641208 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.645256 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmkf8\" (UniqueName: \"kubernetes.io/projected/95f16522-f695-414f-8ec6-c2cb86913080-kube-api-access-rmkf8\") pod \"cinder-operator-controller-manager-79856dc55c-wp2d7\" (UID: \"95f16522-f695-414f-8ec6-c2cb86913080\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.645353 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xkj5\" (UniqueName: \"kubernetes.io/projected/f779482e-56ae-49de-8d6d-e06bf57dc3a7-kube-api-access-8xkj5\") pod \"barbican-operator-controller-manager-86dc4d89c8-6rjkn\" (UID: \"f779482e-56ae-49de-8d6d-e06bf57dc3a7\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.645420 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znbgp\" (UniqueName: \"kubernetes.io/projected/f4b3bf02-333f-46c2-90a0-fe92b166328c-kube-api-access-znbgp\") pod \"designate-operator-controller-manager-7d695c9b56-p8vr6\" (UID: \"f4b3bf02-333f-46c2-90a0-fe92b166328c\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.655348 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.656261 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.659337 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-vqgsv" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.673330 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.674558 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.676201 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-45hlx" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.691503 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.699524 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.741912 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.742849 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.748132 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hjbvb" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.748837 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfvn2\" (UniqueName: \"kubernetes.io/projected/d99686fc-e38b-49de-89a3-7de31352efc3-kube-api-access-gfvn2\") pod \"glance-operator-controller-manager-68b95954c9-ws7tg\" (UID: \"d99686fc-e38b-49de-89a3-7de31352efc3\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.748865 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xkj5\" (UniqueName: \"kubernetes.io/projected/f779482e-56ae-49de-8d6d-e06bf57dc3a7-kube-api-access-8xkj5\") pod \"barbican-operator-controller-manager-86dc4d89c8-6rjkn\" (UID: \"f779482e-56ae-49de-8d6d-e06bf57dc3a7\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.748909 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22w99\" (UniqueName: \"kubernetes.io/projected/9ae5fef4-4b29-4ea2-8404-b145d960ef4a-kube-api-access-22w99\") pod \"heat-operator-controller-manager-774b86978c-ckj7f\" (UID: \"9ae5fef4-4b29-4ea2-8404-b145d960ef4a\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.748932 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znbgp\" (UniqueName: \"kubernetes.io/projected/f4b3bf02-333f-46c2-90a0-fe92b166328c-kube-api-access-znbgp\") pod \"designate-operator-controller-manager-7d695c9b56-p8vr6\" (UID: \"f4b3bf02-333f-46c2-90a0-fe92b166328c\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.748973 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmkf8\" (UniqueName: \"kubernetes.io/projected/95f16522-f695-414f-8ec6-c2cb86913080-kube-api-access-rmkf8\") pod \"cinder-operator-controller-manager-79856dc55c-wp2d7\" (UID: \"95f16522-f695-414f-8ec6-c2cb86913080\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.755523 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.756710 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.759838 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.760099 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-zgjrh" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.771136 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znbgp\" (UniqueName: \"kubernetes.io/projected/f4b3bf02-333f-46c2-90a0-fe92b166328c-kube-api-access-znbgp\") pod \"designate-operator-controller-manager-7d695c9b56-p8vr6\" (UID: \"f4b3bf02-333f-46c2-90a0-fe92b166328c\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.779573 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmkf8\" (UniqueName: \"kubernetes.io/projected/95f16522-f695-414f-8ec6-c2cb86913080-kube-api-access-rmkf8\") pod \"cinder-operator-controller-manager-79856dc55c-wp2d7\" (UID: \"95f16522-f695-414f-8ec6-c2cb86913080\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.780341 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.813472 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.814162 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xkj5\" (UniqueName: \"kubernetes.io/projected/f779482e-56ae-49de-8d6d-e06bf57dc3a7-kube-api-access-8xkj5\") pod \"barbican-operator-controller-manager-86dc4d89c8-6rjkn\" (UID: \"f779482e-56ae-49de-8d6d-e06bf57dc3a7\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.830950 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.831944 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.845516 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.845569 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.846440 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.846945 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-t6x9x" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.848757 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nh76p" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.850531 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-bvfmn\" (UID: \"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.850576 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22w99\" (UniqueName: \"kubernetes.io/projected/9ae5fef4-4b29-4ea2-8404-b145d960ef4a-kube-api-access-22w99\") pod \"heat-operator-controller-manager-774b86978c-ckj7f\" (UID: \"9ae5fef4-4b29-4ea2-8404-b145d960ef4a\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.850632 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtzbl\" (UniqueName: \"kubernetes.io/projected/0f776395-f2dd-41cc-af5f-e38dcd9da7b3-kube-api-access-wtzbl\") pod \"horizon-operator-controller-manager-68c9694994-mq8nd\" (UID: \"0f776395-f2dd-41cc-af5f-e38dcd9da7b3\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.850667 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfvn2\" (UniqueName: \"kubernetes.io/projected/d99686fc-e38b-49de-89a3-7de31352efc3-kube-api-access-gfvn2\") pod \"glance-operator-controller-manager-68b95954c9-ws7tg\" (UID: \"d99686fc-e38b-49de-89a3-7de31352efc3\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.850686 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwpm9\" (UniqueName: \"kubernetes.io/projected/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-kube-api-access-rwpm9\") pod \"infra-operator-controller-manager-d5cc86f4b-bvfmn\" (UID: \"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.856187 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.857112 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.861097 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-826jg" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.871390 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.872531 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.882457 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rfrxt" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.887280 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.888350 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" Nov 24 19:30:00 crc kubenswrapper[5035]: W1124 19:30:00.889826 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3e46e48_d877_4332_918a_95d8c6afc3b5.slice/crio-33ca944d3845b3f2bc683635e917d7e35696480a05d7077e0ad1dd6afeff1f7e WatchSource:0}: Error finding container 33ca944d3845b3f2bc683635e917d7e35696480a05d7077e0ad1dd6afeff1f7e: Status 404 returned error can't find the container with id 33ca944d3845b3f2bc683635e917d7e35696480a05d7077e0ad1dd6afeff1f7e Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.895400 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-28wd2" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.898626 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22w99\" (UniqueName: \"kubernetes.io/projected/9ae5fef4-4b29-4ea2-8404-b145d960ef4a-kube-api-access-22w99\") pod \"heat-operator-controller-manager-774b86978c-ckj7f\" (UID: \"9ae5fef4-4b29-4ea2-8404-b145d960ef4a\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.905414 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.912195 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.920771 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.925415 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.935848 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfvn2\" (UniqueName: \"kubernetes.io/projected/d99686fc-e38b-49de-89a3-7de31352efc3-kube-api-access-gfvn2\") pod \"glance-operator-controller-manager-68b95954c9-ws7tg\" (UID: \"d99686fc-e38b-49de-89a3-7de31352efc3\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.939142 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.946079 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.946973 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.951468 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbd89\" (UniqueName: \"kubernetes.io/projected/6084f8a1-4cde-4714-af8d-76f274dd4b68-kube-api-access-qbd89\") pod \"ironic-operator-controller-manager-5bfcdc958c-rd2qm\" (UID: \"6084f8a1-4cde-4714-af8d-76f274dd4b68\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.951632 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtzbl\" (UniqueName: \"kubernetes.io/projected/0f776395-f2dd-41cc-af5f-e38dcd9da7b3-kube-api-access-wtzbl\") pod \"horizon-operator-controller-manager-68c9694994-mq8nd\" (UID: \"0f776395-f2dd-41cc-af5f-e38dcd9da7b3\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.951734 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tv96\" (UniqueName: \"kubernetes.io/projected/f835e1e6-2181-4243-9572-0566b72f2f11-kube-api-access-7tv96\") pod \"manila-operator-controller-manager-58bb8d67cc-ln7hm\" (UID: \"f835e1e6-2181-4243-9572-0566b72f2f11\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.951860 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwpm9\" (UniqueName: \"kubernetes.io/projected/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-kube-api-access-rwpm9\") pod \"infra-operator-controller-manager-d5cc86f4b-bvfmn\" (UID: \"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.951965 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzlgv\" (UniqueName: \"kubernetes.io/projected/af7760bb-5727-4a40-b9a8-9c7b48a9d066-kube-api-access-tzlgv\") pod \"keystone-operator-controller-manager-748dc6576f-tfjff\" (UID: \"af7760bb-5727-4a40-b9a8-9c7b48a9d066\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.952100 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb65q\" (UniqueName: \"kubernetes.io/projected/792f458f-945a-4846-ac66-c3e2801c2be6-kube-api-access-gb65q\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-csfwz\" (UID: \"792f458f-945a-4846-ac66-c3e2801c2be6\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.952198 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-bvfmn\" (UID: \"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.952318 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tb79\" (UniqueName: \"kubernetes.io/projected/6db4802c-aa26-4fc1-baa7-37ac2845a460-kube-api-access-6tb79\") pod \"neutron-operator-controller-manager-7c57c8bbc4-5wtlz\" (UID: \"6db4802c-aa26-4fc1-baa7-37ac2845a460\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" Nov 24 19:30:00 crc kubenswrapper[5035]: E1124 19:30:00.952582 5035 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 19:30:00 crc kubenswrapper[5035]: E1124 19:30:00.952717 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-cert podName:c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f nodeName:}" failed. No retries permitted until 2025-11-24 19:30:01.452692339 +0000 UTC m=+939.975198666 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-cert") pod "infra-operator-controller-manager-d5cc86f4b-bvfmn" (UID: "c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f") : secret "infra-operator-webhook-server-cert" not found Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.962723 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-h9mqs" Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.968324 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8"] Nov 24 19:30:00 crc kubenswrapper[5035]: I1124 19:30:00.969232 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.027533 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.028123 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.028604 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.040938 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-42zs4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.041843 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.117466 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwpm9\" (UniqueName: \"kubernetes.io/projected/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-kube-api-access-rwpm9\") pod \"infra-operator-controller-manager-d5cc86f4b-bvfmn\" (UID: \"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.118262 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtzbl\" (UniqueName: \"kubernetes.io/projected/0f776395-f2dd-41cc-af5f-e38dcd9da7b3-kube-api-access-wtzbl\") pod \"horizon-operator-controller-manager-68c9694994-mq8nd\" (UID: \"0f776395-f2dd-41cc-af5f-e38dcd9da7b3\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.120724 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.121199 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzlgv\" (UniqueName: \"kubernetes.io/projected/af7760bb-5727-4a40-b9a8-9c7b48a9d066-kube-api-access-tzlgv\") pod \"keystone-operator-controller-manager-748dc6576f-tfjff\" (UID: \"af7760bb-5727-4a40-b9a8-9c7b48a9d066\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.121242 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb65q\" (UniqueName: \"kubernetes.io/projected/792f458f-945a-4846-ac66-c3e2801c2be6-kube-api-access-gb65q\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-csfwz\" (UID: \"792f458f-945a-4846-ac66-c3e2801c2be6\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.122034 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tb79\" (UniqueName: \"kubernetes.io/projected/6db4802c-aa26-4fc1-baa7-37ac2845a460-kube-api-access-6tb79\") pod \"neutron-operator-controller-manager-7c57c8bbc4-5wtlz\" (UID: \"6db4802c-aa26-4fc1-baa7-37ac2845a460\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.122074 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrkzp\" (UniqueName: \"kubernetes.io/projected/00919c94-bf31-4ea8-a322-5b81bdb050c9-kube-api-access-vrkzp\") pod \"octavia-operator-controller-manager-fd75fd47d-rqf69\" (UID: \"00919c94-bf31-4ea8-a322-5b81bdb050c9\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.122118 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs4bc\" (UniqueName: \"kubernetes.io/projected/bf589985-cf9b-4133-8d16-794c8fbfa0f3-kube-api-access-xs4bc\") pod \"nova-operator-controller-manager-79556f57fc-kkcw8\" (UID: \"bf589985-cf9b-4133-8d16-794c8fbfa0f3\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.123037 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbd89\" (UniqueName: \"kubernetes.io/projected/6084f8a1-4cde-4714-af8d-76f274dd4b68-kube-api-access-qbd89\") pod \"ironic-operator-controller-manager-5bfcdc958c-rd2qm\" (UID: \"6084f8a1-4cde-4714-af8d-76f274dd4b68\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.123112 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tv96\" (UniqueName: \"kubernetes.io/projected/f835e1e6-2181-4243-9572-0566b72f2f11-kube-api-access-7tv96\") pod \"manila-operator-controller-manager-58bb8d67cc-ln7hm\" (UID: \"f835e1e6-2181-4243-9572-0566b72f2f11\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.149915 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.154633 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.162186 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzlgv\" (UniqueName: \"kubernetes.io/projected/af7760bb-5727-4a40-b9a8-9c7b48a9d066-kube-api-access-tzlgv\") pod \"keystone-operator-controller-manager-748dc6576f-tfjff\" (UID: \"af7760bb-5727-4a40-b9a8-9c7b48a9d066\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.168255 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb65q\" (UniqueName: \"kubernetes.io/projected/792f458f-945a-4846-ac66-c3e2801c2be6-kube-api-access-gb65q\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-csfwz\" (UID: \"792f458f-945a-4846-ac66-c3e2801c2be6\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.169678 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbd89\" (UniqueName: \"kubernetes.io/projected/6084f8a1-4cde-4714-af8d-76f274dd4b68-kube-api-access-qbd89\") pod \"ironic-operator-controller-manager-5bfcdc958c-rd2qm\" (UID: \"6084f8a1-4cde-4714-af8d-76f274dd4b68\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.169987 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.187516 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tb79\" (UniqueName: \"kubernetes.io/projected/6db4802c-aa26-4fc1-baa7-37ac2845a460-kube-api-access-6tb79\") pod \"neutron-operator-controller-manager-7c57c8bbc4-5wtlz\" (UID: \"6db4802c-aa26-4fc1-baa7-37ac2845a460\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.196031 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.197507 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tv96\" (UniqueName: \"kubernetes.io/projected/f835e1e6-2181-4243-9572-0566b72f2f11-kube-api-access-7tv96\") pod \"manila-operator-controller-manager-58bb8d67cc-ln7hm\" (UID: \"f835e1e6-2181-4243-9572-0566b72f2f11\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.212718 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.216450 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.218326 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.221386 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-kn4rc" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.226223 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbdbj\" (UniqueName: \"kubernetes.io/projected/ba2f4f9d-09b7-4192-9061-39d5204dd054-kube-api-access-gbdbj\") pod \"ovn-operator-controller-manager-66cf5c67ff-prr8k\" (UID: \"ba2f4f9d-09b7-4192-9061-39d5204dd054\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.226520 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrkzp\" (UniqueName: \"kubernetes.io/projected/00919c94-bf31-4ea8-a322-5b81bdb050c9-kube-api-access-vrkzp\") pod \"octavia-operator-controller-manager-fd75fd47d-rqf69\" (UID: \"00919c94-bf31-4ea8-a322-5b81bdb050c9\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.226602 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs4bc\" (UniqueName: \"kubernetes.io/projected/bf589985-cf9b-4133-8d16-794c8fbfa0f3-kube-api-access-xs4bc\") pod \"nova-operator-controller-manager-79556f57fc-kkcw8\" (UID: \"bf589985-cf9b-4133-8d16-794c8fbfa0f3\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.229314 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.233776 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.233899 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.238554 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-px272" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.265031 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrkzp\" (UniqueName: \"kubernetes.io/projected/00919c94-bf31-4ea8-a322-5b81bdb050c9-kube-api-access-vrkzp\") pod \"octavia-operator-controller-manager-fd75fd47d-rqf69\" (UID: \"00919c94-bf31-4ea8-a322-5b81bdb050c9\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.265556 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.266013 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.267189 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.273519 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.274606 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-v6rn9" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.274785 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.284460 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.289599 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.303363 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs4bc\" (UniqueName: \"kubernetes.io/projected/bf589985-cf9b-4133-8d16-794c8fbfa0f3-kube-api-access-xs4bc\") pod \"nova-operator-controller-manager-79556f57fc-kkcw8\" (UID: \"bf589985-cf9b-4133-8d16-794c8fbfa0f3\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.312402 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.313625 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.321783 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9js2x" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.327645 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbdbj\" (UniqueName: \"kubernetes.io/projected/ba2f4f9d-09b7-4192-9061-39d5204dd054-kube-api-access-gbdbj\") pod \"ovn-operator-controller-manager-66cf5c67ff-prr8k\" (UID: \"ba2f4f9d-09b7-4192-9061-39d5204dd054\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.328798 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.340841 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.341117 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.345010 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-rh9jx" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.350425 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.357681 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.361688 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.362916 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.368599 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.374414 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbdbj\" (UniqueName: \"kubernetes.io/projected/ba2f4f9d-09b7-4192-9061-39d5204dd054-kube-api-access-gbdbj\") pod \"ovn-operator-controller-manager-66cf5c67ff-prr8k\" (UID: \"ba2f4f9d-09b7-4192-9061-39d5204dd054\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.387621 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-mptdw" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.390507 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-rjs7d"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.395621 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.398024 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mvksd" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.398223 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-rjs7d"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.430808 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtnlc\" (UniqueName: \"kubernetes.io/projected/8ba349a2-1cc6-43ef-9045-4d02feffffd9-kube-api-access-jtnlc\") pod \"placement-operator-controller-manager-5db546f9d9-87rk8\" (UID: \"8ba349a2-1cc6-43ef-9045-4d02feffffd9\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.430895 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg6vv\" (UniqueName: \"kubernetes.io/projected/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-kube-api-access-pg6vv\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.430924 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvthh\" (UniqueName: \"kubernetes.io/projected/3d6e2a89-1911-4c26-8a9c-03c9955af913-kube-api-access-cvthh\") pod \"watcher-operator-controller-manager-864885998-rjs7d\" (UID: \"3d6e2a89-1911-4c26-8a9c-03c9955af913\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.430969 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.430993 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzmfk\" (UniqueName: \"kubernetes.io/projected/be99a698-bf30-4dd1-b534-e785a37143bc-kube-api-access-qzmfk\") pod \"swift-operator-controller-manager-6fdc4fcf86-tpw7d\" (UID: \"be99a698-bf30-4dd1-b534-e785a37143bc\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.431017 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l24pf\" (UniqueName: \"kubernetes.io/projected/5d0cb726-bb99-44fe-97c5-000e9619a714-kube-api-access-l24pf\") pod \"test-operator-controller-manager-5cb74df96-vjkfm\" (UID: \"5d0cb726-bb99-44fe-97c5-000e9619a714\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.431043 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w424h\" (UniqueName: \"kubernetes.io/projected/eab64b25-1263-4d64-ba86-1f7227feae79-kube-api-access-w424h\") pod \"telemetry-operator-controller-manager-567f98c9d-grqcw\" (UID: \"eab64b25-1263-4d64-ba86-1f7227feae79\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.467504 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.499646 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.513572 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.514576 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.520256 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-66z6g" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.520446 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.520598 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532365 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532403 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzmfk\" (UniqueName: \"kubernetes.io/projected/be99a698-bf30-4dd1-b534-e785a37143bc-kube-api-access-qzmfk\") pod \"swift-operator-controller-manager-6fdc4fcf86-tpw7d\" (UID: \"be99a698-bf30-4dd1-b534-e785a37143bc\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532427 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l24pf\" (UniqueName: \"kubernetes.io/projected/5d0cb726-bb99-44fe-97c5-000e9619a714-kube-api-access-l24pf\") pod \"test-operator-controller-manager-5cb74df96-vjkfm\" (UID: \"5d0cb726-bb99-44fe-97c5-000e9619a714\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532480 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w424h\" (UniqueName: \"kubernetes.io/projected/eab64b25-1263-4d64-ba86-1f7227feae79-kube-api-access-w424h\") pod \"telemetry-operator-controller-manager-567f98c9d-grqcw\" (UID: \"eab64b25-1263-4d64-ba86-1f7227feae79\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532558 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr7zv\" (UniqueName: \"kubernetes.io/projected/4eb72ce7-5f41-4360-91be-e7bce1db7b69-kube-api-access-xr7zv\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532612 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-bvfmn\" (UID: \"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532636 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532688 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtnlc\" (UniqueName: \"kubernetes.io/projected/8ba349a2-1cc6-43ef-9045-4d02feffffd9-kube-api-access-jtnlc\") pod \"placement-operator-controller-manager-5db546f9d9-87rk8\" (UID: \"8ba349a2-1cc6-43ef-9045-4d02feffffd9\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.532730 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.533811 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg6vv\" (UniqueName: \"kubernetes.io/projected/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-kube-api-access-pg6vv\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.533841 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvthh\" (UniqueName: \"kubernetes.io/projected/3d6e2a89-1911-4c26-8a9c-03c9955af913-kube-api-access-cvthh\") pod \"watcher-operator-controller-manager-864885998-rjs7d\" (UID: \"3d6e2a89-1911-4c26-8a9c-03c9955af913\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" Nov 24 19:30:01 crc kubenswrapper[5035]: E1124 19:30:01.534375 5035 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 19:30:01 crc kubenswrapper[5035]: E1124 19:30:01.534433 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert podName:b10f4358-81d9-4d88-a0c5-3aa698f00cdf nodeName:}" failed. No retries permitted until 2025-11-24 19:30:02.034417873 +0000 UTC m=+940.556924130 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert") pod "openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" (UID: "b10f4358-81d9-4d88-a0c5-3aa698f00cdf") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.543704 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.543943 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-bvfmn\" (UID: \"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.560527 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvthh\" (UniqueName: \"kubernetes.io/projected/3d6e2a89-1911-4c26-8a9c-03c9955af913-kube-api-access-cvthh\") pod \"watcher-operator-controller-manager-864885998-rjs7d\" (UID: \"3d6e2a89-1911-4c26-8a9c-03c9955af913\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.564015 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzmfk\" (UniqueName: \"kubernetes.io/projected/be99a698-bf30-4dd1-b534-e785a37143bc-kube-api-access-qzmfk\") pod \"swift-operator-controller-manager-6fdc4fcf86-tpw7d\" (UID: \"be99a698-bf30-4dd1-b534-e785a37143bc\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.567779 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg6vv\" (UniqueName: \"kubernetes.io/projected/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-kube-api-access-pg6vv\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.567806 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtnlc\" (UniqueName: \"kubernetes.io/projected/8ba349a2-1cc6-43ef-9045-4d02feffffd9-kube-api-access-jtnlc\") pod \"placement-operator-controller-manager-5db546f9d9-87rk8\" (UID: \"8ba349a2-1cc6-43ef-9045-4d02feffffd9\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.573634 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.580037 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l24pf\" (UniqueName: \"kubernetes.io/projected/5d0cb726-bb99-44fe-97c5-000e9619a714-kube-api-access-l24pf\") pod \"test-operator-controller-manager-5cb74df96-vjkfm\" (UID: \"5d0cb726-bb99-44fe-97c5-000e9619a714\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.584828 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w424h\" (UniqueName: \"kubernetes.io/projected/eab64b25-1263-4d64-ba86-1f7227feae79-kube-api-access-w424h\") pod \"telemetry-operator-controller-manager-567f98c9d-grqcw\" (UID: \"eab64b25-1263-4d64-ba86-1f7227feae79\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.584907 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.602132 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.612333 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.629115 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.630779 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.631641 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.635397 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr7zv\" (UniqueName: \"kubernetes.io/projected/4eb72ce7-5f41-4360-91be-e7bce1db7b69-kube-api-access-xr7zv\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.635537 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.635602 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: E1124 19:30:01.635763 5035 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 19:30:01 crc kubenswrapper[5035]: E1124 19:30:01.635815 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs podName:4eb72ce7-5f41-4360-91be-e7bce1db7b69 nodeName:}" failed. No retries permitted until 2025-11-24 19:30:02.135797914 +0000 UTC m=+940.658304171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-hvjd4" (UID: "4eb72ce7-5f41-4360-91be-e7bce1db7b69") : secret "webhook-server-cert" not found Nov 24 19:30:01 crc kubenswrapper[5035]: E1124 19:30:01.635866 5035 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 24 19:30:01 crc kubenswrapper[5035]: E1124 19:30:01.635909 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs podName:4eb72ce7-5f41-4360-91be-e7bce1db7b69 nodeName:}" failed. No retries permitted until 2025-11-24 19:30:02.135893657 +0000 UTC m=+940.658399914 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs") pod "openstack-operator-controller-manager-7cd5954d9-hvjd4" (UID: "4eb72ce7-5f41-4360-91be-e7bce1db7b69") : secret "metrics-server-cert" not found Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.644571 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-97w5f" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.659931 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.664802 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr7zv\" (UniqueName: \"kubernetes.io/projected/4eb72ce7-5f41-4360-91be-e7bce1db7b69-kube-api-access-xr7zv\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.676550 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.709361 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn"] Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.731345 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.736660 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnwc9\" (UniqueName: \"kubernetes.io/projected/98c2e05e-f4af-4979-960b-4104a843321a-kube-api-access-wnwc9\") pod \"rabbitmq-cluster-operator-manager-668c99d594-n2s8q\" (UID: \"98c2e05e-f4af-4979-960b-4104a843321a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.798236 5035 generic.go:334] "Generic (PLEG): container finished" podID="b3e46e48-d877-4332-918a-95d8c6afc3b5" containerID="7a4970c6567426499877620ccac9fa6bcb84b68ebdfaaab1c0abc47abb90c7dc" exitCode=0 Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.798281 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" event={"ID":"b3e46e48-d877-4332-918a-95d8c6afc3b5","Type":"ContainerDied","Data":"7a4970c6567426499877620ccac9fa6bcb84b68ebdfaaab1c0abc47abb90c7dc"} Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.798321 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" event={"ID":"b3e46e48-d877-4332-918a-95d8c6afc3b5","Type":"ContainerStarted","Data":"33ca944d3845b3f2bc683635e917d7e35696480a05d7077e0ad1dd6afeff1f7e"} Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.838436 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnwc9\" (UniqueName: \"kubernetes.io/projected/98c2e05e-f4af-4979-960b-4104a843321a-kube-api-access-wnwc9\") pod \"rabbitmq-cluster-operator-manager-668c99d594-n2s8q\" (UID: \"98c2e05e-f4af-4979-960b-4104a843321a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" Nov 24 19:30:01 crc kubenswrapper[5035]: I1124 19:30:01.855783 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnwc9\" (UniqueName: \"kubernetes.io/projected/98c2e05e-f4af-4979-960b-4104a843321a-kube-api-access-wnwc9\") pod \"rabbitmq-cluster-operator-manager-668c99d594-n2s8q\" (UID: \"98c2e05e-f4af-4979-960b-4104a843321a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.041709 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.041907 5035 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.041988 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert podName:b10f4358-81d9-4d88-a0c5-3aa698f00cdf nodeName:}" failed. No retries permitted until 2025-11-24 19:30:03.041954293 +0000 UTC m=+941.564460550 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert") pod "openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" (UID: "b10f4358-81d9-4d88-a0c5-3aa698f00cdf") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.089179 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.144976 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.145401 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.145195 5035 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.145504 5035 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.145519 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs podName:4eb72ce7-5f41-4360-91be-e7bce1db7b69 nodeName:}" failed. No retries permitted until 2025-11-24 19:30:03.145494203 +0000 UTC m=+941.668000460 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-hvjd4" (UID: "4eb72ce7-5f41-4360-91be-e7bce1db7b69") : secret "webhook-server-cert" not found Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.145534 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs podName:4eb72ce7-5f41-4360-91be-e7bce1db7b69 nodeName:}" failed. No retries permitted until 2025-11-24 19:30:03.145526644 +0000 UTC m=+941.668032901 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs") pod "openstack-operator-controller-manager-7cd5954d9-hvjd4" (UID: "4eb72ce7-5f41-4360-91be-e7bce1db7b69") : secret "metrics-server-cert" not found Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.169272 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.175607 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg"] Nov 24 19:30:02 crc kubenswrapper[5035]: W1124 19:30:02.309750 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95f16522_f695_414f_8ec6_c2cb86913080.slice/crio-a6ba99675a66126fd06c3c94e8b982d41fb6dcc0cb7fffbf118424892c9d6522 WatchSource:0}: Error finding container a6ba99675a66126fd06c3c94e8b982d41fb6dcc0cb7fffbf118424892c9d6522: Status 404 returned error can't find the container with id a6ba99675a66126fd06c3c94e8b982d41fb6dcc0cb7fffbf118424892c9d6522 Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.310449 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7"] Nov 24 19:30:02 crc kubenswrapper[5035]: W1124 19:30:02.319370 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6084f8a1_4cde_4714_af8d_76f274dd4b68.slice/crio-8c6effdaaf684acc42f0b83eba530bd723accf23970bd5f28b76146140149b9f WatchSource:0}: Error finding container 8c6effdaaf684acc42f0b83eba530bd723accf23970bd5f28b76146140149b9f: Status 404 returned error can't find the container with id 8c6effdaaf684acc42f0b83eba530bd723accf23970bd5f28b76146140149b9f Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.321584 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.340044 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm"] Nov 24 19:30:02 crc kubenswrapper[5035]: W1124 19:30:02.343122 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf835e1e6_2181_4243_9572_0566b72f2f11.slice/crio-de172ac1bc4b996013a0e253ee456c017a753aab2219ed6c22fa3d19ed27e7e6 WatchSource:0}: Error finding container de172ac1bc4b996013a0e253ee456c017a753aab2219ed6c22fa3d19ed27e7e6: Status 404 returned error can't find the container with id de172ac1bc4b996013a0e253ee456c017a753aab2219ed6c22fa3d19ed27e7e6 Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.537275 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.547957 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.555370 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.559773 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.566382 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.575425 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.738874 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.748780 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.754907 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-rjs7d"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.779525 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm"] Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.788378 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8"] Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.791739 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cvthh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-rjs7d_openstack-operators(3d6e2a89-1911-4c26-8a9c-03c9955af913): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.793191 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69"] Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.793165 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbdbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-prr8k_openstack-operators(ba2f4f9d-09b7-4192-9061-39d5204dd054): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.795137 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cvthh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-rjs7d_openstack-operators(3d6e2a89-1911-4c26-8a9c-03c9955af913): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.796509 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" podUID="3d6e2a89-1911-4c26-8a9c-03c9955af913" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.797422 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbdbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-prr8k_openstack-operators(ba2f4f9d-09b7-4192-9061-39d5204dd054): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.798533 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" podUID="ba2f4f9d-09b7-4192-9061-39d5204dd054" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.802522 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jtnlc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5db546f9d9-87rk8_openstack-operators(8ba349a2-1cc6-43ef-9045-4d02feffffd9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.802770 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qzmfk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-tpw7d_openstack-operators(be99a698-bf30-4dd1-b534-e785a37143bc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.803276 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d"] Nov 24 19:30:02 crc kubenswrapper[5035]: W1124 19:30:02.805806 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98c2e05e_f4af_4979_960b_4104a843321a.slice/crio-bd00218f515fc6c7d822fe89170828ab852ca6a042254ceaddde1af589681d4a WatchSource:0}: Error finding container bd00218f515fc6c7d822fe89170828ab852ca6a042254ceaddde1af589681d4a: Status 404 returned error can't find the container with id bd00218f515fc6c7d822fe89170828ab852ca6a042254ceaddde1af589681d4a Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.806414 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" event={"ID":"0f776395-f2dd-41cc-af5f-e38dcd9da7b3","Type":"ContainerStarted","Data":"5c140fa69a6ccfbfdf4803b8acd169a4700ea261af0017267859c76246484352"} Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.807965 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qzmfk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-tpw7d_openstack-operators(be99a698-bf30-4dd1-b534-e785a37143bc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.808185 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jtnlc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5db546f9d9-87rk8_openstack-operators(8ba349a2-1cc6-43ef-9045-4d02feffffd9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.809061 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" podUID="be99a698-bf30-4dd1-b534-e785a37143bc" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.809450 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wnwc9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-n2s8q_openstack-operators(98c2e05e-f4af-4979-960b-4104a843321a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.809510 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" podUID="8ba349a2-1cc6-43ef-9045-4d02feffffd9" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.810518 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" podUID="98c2e05e-f4af-4979-960b-4104a843321a" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.810587 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" event={"ID":"6db4802c-aa26-4fc1-baa7-37ac2845a460","Type":"ContainerStarted","Data":"b035ad6760fc51733957b9b1197ec7ead1439fea48453a077c10e4c144d6ae2d"} Nov 24 19:30:02 crc kubenswrapper[5035]: W1124 19:30:02.811471 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab64b25_1263_4d64_ba86_1f7227feae79.slice/crio-585c486bb7abd228672da019dc3ec6a11eca82de9c4879792c933d6dca4d29fb WatchSource:0}: Error finding container 585c486bb7abd228672da019dc3ec6a11eca82de9c4879792c933d6dca4d29fb: Status 404 returned error can't find the container with id 585c486bb7abd228672da019dc3ec6a11eca82de9c4879792c933d6dca4d29fb Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.812080 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q"] Nov 24 19:30:02 crc kubenswrapper[5035]: W1124 19:30:02.812673 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d0cb726_bb99_44fe_97c5_000e9619a714.slice/crio-e9fb9da303cabc9e96dcd8513b359358eaab0fa7d64e7a8f4673297c0366eab6 WatchSource:0}: Error finding container e9fb9da303cabc9e96dcd8513b359358eaab0fa7d64e7a8f4673297c0366eab6: Status 404 returned error can't find the container with id e9fb9da303cabc9e96dcd8513b359358eaab0fa7d64e7a8f4673297c0366eab6 Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.813623 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" event={"ID":"ba2f4f9d-09b7-4192-9061-39d5204dd054","Type":"ContainerStarted","Data":"b0524c96305eafff0472bf49d16959aa58053292eff69a51811793da83cf183e"} Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.816483 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w424h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-grqcw_openstack-operators(eab64b25-1263-4d64-ba86-1f7227feae79): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.817121 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vrkzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-rqf69_openstack-operators(00919c94-bf31-4ea8-a322-5b81bdb050c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.817124 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l24pf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-vjkfm_openstack-operators(5d0cb726-bb99-44fe-97c5-000e9619a714): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.817430 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw"] Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.817861 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" podUID="ba2f4f9d-09b7-4192-9061-39d5204dd054" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.818481 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w424h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-grqcw_openstack-operators(eab64b25-1263-4d64-ba86-1f7227feae79): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.819395 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vrkzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-rqf69_openstack-operators(00919c94-bf31-4ea8-a322-5b81bdb050c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.819480 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l24pf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-vjkfm_openstack-operators(5d0cb726-bb99-44fe-97c5-000e9619a714): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.819900 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" podUID="eab64b25-1263-4d64-ba86-1f7227feae79" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.820227 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" event={"ID":"f835e1e6-2181-4243-9572-0566b72f2f11","Type":"ContainerStarted","Data":"de172ac1bc4b996013a0e253ee456c017a753aab2219ed6c22fa3d19ed27e7e6"} Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.820812 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.821180 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" podUID="00919c94-bf31-4ea8-a322-5b81bdb050c9" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.822239 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" event={"ID":"f4b3bf02-333f-46c2-90a0-fe92b166328c","Type":"ContainerStarted","Data":"d9b8c9d17e133bd89184d360b736029a9f7473c6ae5b2402c3c1e911932bf851"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.824027 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" event={"ID":"f779482e-56ae-49de-8d6d-e06bf57dc3a7","Type":"ContainerStarted","Data":"09342cdc86d1b12ecfa5c54c1b0236d7a0ba8ecb5b149385b53a4b034fc7085c"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.825113 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" event={"ID":"bf589985-cf9b-4133-8d16-794c8fbfa0f3","Type":"ContainerStarted","Data":"5f13b237f5424b44b3636e13c3c2b2bf3ed43b875950f6751d5491dde3c9a2c3"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.832340 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" event={"ID":"6084f8a1-4cde-4714-af8d-76f274dd4b68","Type":"ContainerStarted","Data":"8c6effdaaf684acc42f0b83eba530bd723accf23970bd5f28b76146140149b9f"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.833624 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" event={"ID":"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f","Type":"ContainerStarted","Data":"748a1684c6a6baf7042f0763b01eb98d7064aeb0f07be2927cd56253ba28d0dd"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.836644 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" event={"ID":"95f16522-f695-414f-8ec6-c2cb86913080","Type":"ContainerStarted","Data":"a6ba99675a66126fd06c3c94e8b982d41fb6dcc0cb7fffbf118424892c9d6522"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.838208 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" event={"ID":"3d6e2a89-1911-4c26-8a9c-03c9955af913","Type":"ContainerStarted","Data":"be6d1073f00ba7faecc6193386553690e5980c346132f58fe9e241be0cad31d7"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.840150 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" event={"ID":"9ae5fef4-4b29-4ea2-8404-b145d960ef4a","Type":"ContainerStarted","Data":"493535b01f73e192b50dbab4e253fff91d9f0901d450e62de1ebb00268ad97df"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.841516 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" event={"ID":"be99a698-bf30-4dd1-b534-e785a37143bc","Type":"ContainerStarted","Data":"abca4489de5efac1edb0e9333da25192b9a39969e1351dbf7bbd7efbfacd763a"} Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.842126 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" podUID="3d6e2a89-1911-4c26-8a9c-03c9955af913" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.842782 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" event={"ID":"af7760bb-5727-4a40-b9a8-9c7b48a9d066","Type":"ContainerStarted","Data":"51b622ad7bc04e8584b10697335797934e02ff8c835aa2859c864350512d848f"} Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.843128 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" podUID="be99a698-bf30-4dd1-b534-e785a37143bc" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.846361 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" event={"ID":"8ba349a2-1cc6-43ef-9045-4d02feffffd9","Type":"ContainerStarted","Data":"9d64e505ffdd6b83c1e319618a214ce9d02079bad726527cd90d807052b0c6b6"} Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.852090 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" event={"ID":"d99686fc-e38b-49de-89a3-7de31352efc3","Type":"ContainerStarted","Data":"2a06fe332f8b4919dc098262a2073b5038a81f094e2ad747dd20d610c239c87d"} Nov 24 19:30:02 crc kubenswrapper[5035]: E1124 19:30:02.857938 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" podUID="8ba349a2-1cc6-43ef-9045-4d02feffffd9" Nov 24 19:30:02 crc kubenswrapper[5035]: I1124 19:30:02.860051 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" event={"ID":"792f458f-945a-4846-ac66-c3e2801c2be6","Type":"ContainerStarted","Data":"30257ac495f537773fc41ecaefd86f3665347f2742bff686917b935bae170d56"} Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.067060 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.078764 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b10f4358-81d9-4d88-a0c5-3aa698f00cdf-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gqwwm\" (UID: \"b10f4358-81d9-4d88-a0c5-3aa698f00cdf\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.104705 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.115146 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.168371 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.168469 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.168672 5035 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.168736 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs podName:4eb72ce7-5f41-4360-91be-e7bce1db7b69 nodeName:}" failed. No retries permitted until 2025-11-24 19:30:05.168718377 +0000 UTC m=+943.691224634 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-hvjd4" (UID: "4eb72ce7-5f41-4360-91be-e7bce1db7b69") : secret "webhook-server-cert" not found Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.169084 5035 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.169119 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs podName:4eb72ce7-5f41-4360-91be-e7bce1db7b69 nodeName:}" failed. No retries permitted until 2025-11-24 19:30:05.169108527 +0000 UTC m=+943.691614784 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs") pod "openstack-operator-controller-manager-7cd5954d9-hvjd4" (UID: "4eb72ce7-5f41-4360-91be-e7bce1db7b69") : secret "metrics-server-cert" not found Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.269104 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e46e48-d877-4332-918a-95d8c6afc3b5-secret-volume\") pod \"b3e46e48-d877-4332-918a-95d8c6afc3b5\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.269283 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e46e48-d877-4332-918a-95d8c6afc3b5-config-volume\") pod \"b3e46e48-d877-4332-918a-95d8c6afc3b5\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.269346 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7lnl\" (UniqueName: \"kubernetes.io/projected/b3e46e48-d877-4332-918a-95d8c6afc3b5-kube-api-access-g7lnl\") pod \"b3e46e48-d877-4332-918a-95d8c6afc3b5\" (UID: \"b3e46e48-d877-4332-918a-95d8c6afc3b5\") " Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.272156 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e46e48-d877-4332-918a-95d8c6afc3b5-config-volume" (OuterVolumeSpecName: "config-volume") pod "b3e46e48-d877-4332-918a-95d8c6afc3b5" (UID: "b3e46e48-d877-4332-918a-95d8c6afc3b5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.274012 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e46e48-d877-4332-918a-95d8c6afc3b5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b3e46e48-d877-4332-918a-95d8c6afc3b5" (UID: "b3e46e48-d877-4332-918a-95d8c6afc3b5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.275538 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e46e48-d877-4332-918a-95d8c6afc3b5-kube-api-access-g7lnl" (OuterVolumeSpecName: "kube-api-access-g7lnl") pod "b3e46e48-d877-4332-918a-95d8c6afc3b5" (UID: "b3e46e48-d877-4332-918a-95d8c6afc3b5"). InnerVolumeSpecName "kube-api-access-g7lnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.371309 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3e46e48-d877-4332-918a-95d8c6afc3b5-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.371344 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3e46e48-d877-4332-918a-95d8c6afc3b5-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.371355 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7lnl\" (UniqueName: \"kubernetes.io/projected/b3e46e48-d877-4332-918a-95d8c6afc3b5-kube-api-access-g7lnl\") on node \"crc\" DevicePath \"\"" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.683125 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm"] Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.868343 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" event={"ID":"98c2e05e-f4af-4979-960b-4104a843321a","Type":"ContainerStarted","Data":"bd00218f515fc6c7d822fe89170828ab852ca6a042254ceaddde1af589681d4a"} Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.869928 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" event={"ID":"eab64b25-1263-4d64-ba86-1f7227feae79","Type":"ContainerStarted","Data":"585c486bb7abd228672da019dc3ec6a11eca82de9c4879792c933d6dca4d29fb"} Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.870105 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" podUID="98c2e05e-f4af-4979-960b-4104a843321a" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.872474 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" event={"ID":"00919c94-bf31-4ea8-a322-5b81bdb050c9","Type":"ContainerStarted","Data":"a3b943e2a87c38fbe281b6a7a4255d0bcccfd599cfd1ad58e9e3aee48c7055da"} Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.875326 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" podUID="00919c94-bf31-4ea8-a322-5b81bdb050c9" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.875371 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" event={"ID":"b10f4358-81d9-4d88-a0c5-3aa698f00cdf","Type":"ContainerStarted","Data":"02373952e8373ac1c859b43ed9ca75ec2a9589de034199d25df9e214c82e8bd3"} Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.878440 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" event={"ID":"b3e46e48-d877-4332-918a-95d8c6afc3b5","Type":"ContainerDied","Data":"33ca944d3845b3f2bc683635e917d7e35696480a05d7077e0ad1dd6afeff1f7e"} Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.878511 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33ca944d3845b3f2bc683635e917d7e35696480a05d7077e0ad1dd6afeff1f7e" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.878598 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w" Nov 24 19:30:03 crc kubenswrapper[5035]: I1124 19:30:03.881037 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" event={"ID":"5d0cb726-bb99-44fe-97c5-000e9619a714","Type":"ContainerStarted","Data":"e9fb9da303cabc9e96dcd8513b359358eaab0fa7d64e7a8f4673297c0366eab6"} Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.893026 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" podUID="eab64b25-1263-4d64-ba86-1f7227feae79" Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.895572 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" podUID="ba2f4f9d-09b7-4192-9061-39d5204dd054" Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.895617 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" podUID="be99a698-bf30-4dd1-b534-e785a37143bc" Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.896064 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" podUID="3d6e2a89-1911-4c26-8a9c-03c9955af913" Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.896685 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" Nov 24 19:30:03 crc kubenswrapper[5035]: E1124 19:30:03.897359 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" podUID="8ba349a2-1cc6-43ef-9045-4d02feffffd9" Nov 24 19:30:04 crc kubenswrapper[5035]: E1124 19:30:04.888974 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" podUID="98c2e05e-f4af-4979-960b-4104a843321a" Nov 24 19:30:04 crc kubenswrapper[5035]: E1124 19:30:04.889992 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" podUID="00919c94-bf31-4ea8-a322-5b81bdb050c9" Nov 24 19:30:04 crc kubenswrapper[5035]: E1124 19:30:04.890100 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" Nov 24 19:30:04 crc kubenswrapper[5035]: E1124 19:30:04.895496 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" podUID="eab64b25-1263-4d64-ba86-1f7227feae79" Nov 24 19:30:05 crc kubenswrapper[5035]: I1124 19:30:05.211774 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:05 crc kubenswrapper[5035]: I1124 19:30:05.212141 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:05 crc kubenswrapper[5035]: I1124 19:30:05.217727 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:05 crc kubenswrapper[5035]: I1124 19:30:05.228363 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4eb72ce7-5f41-4360-91be-e7bce1db7b69-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-hvjd4\" (UID: \"4eb72ce7-5f41-4360-91be-e7bce1db7b69\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:05 crc kubenswrapper[5035]: I1124 19:30:05.262563 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:14 crc kubenswrapper[5035]: E1124 19:30:14.683462 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96" Nov 24 19:30:14 crc kubenswrapper[5035]: E1124 19:30:14.684513 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:5edd825a235f5784d9a65892763c5388c39df1731d0fcbf4ee33408b8c83ac96,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-22w99,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-774b86978c-ckj7f_openstack-operators(9ae5fef4-4b29-4ea2-8404-b145d960ef4a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:30:15 crc kubenswrapper[5035]: E1124 19:30:15.214828 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd" Nov 24 19:30:15 crc kubenswrapper[5035]: E1124 19:30:15.215309 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:7dbadf7b98f2f305f9f1382f55a084c8ca404f4263f76b28e56bd0dc437e2192,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:c8101c77a82eae4407e41e1fd766dfc6e1b7f9ed1679e3efb6f91ff97a1557b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:eb9743b21bbadca6f7cb9ac4fc46b5d58c51c674073c7e1121f4474a71304071,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:3d81f839b98c2e2a5bf0da79f2f9a92dff7d0a3c5a830b0e95c89dad8cf98a6a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:d19ac99249b47dd8ea16cd6aaa5756346aa8a2f119ee50819c15c5366efb417d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:8536169e5537fe6c330eba814248abdcf39cdd8f7e7336034d74e6fda9544050,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:4f1fa337760e82bfd67cdd142a97c121146dd7e621daac161940dd5e4ddb80dc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:3613b345d5baed98effd906f8b0242d863e14c97078ea473ef01fe1b0afc46f3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:9f9f367ed4c85efb16c3a74a4bb707ff0db271d7bc5abc70a71e984b55f43003,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:b73ad22b4955b06d584bce81742556d8c0c7828c495494f8ea7c99391c61b70f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:aa1d3aaf6b394621ed4089a98e0a82b763f467e8b5c5db772f9fdf99fc86e333,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:d6661053141b6df421288a7c9968a155ab82e478c1d75ab41f2cebe2f0ca02d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:ce2d63258cb4e7d0d1c07234de6889c5434464190906798019311a1c7cf6387f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:0485ef9e5b4437f7cd2ba54034a87722ce4669ee86b3773c6b0c037ed8000e91,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api@sha256:43f8a00cd714c59f2c517fe6fabb63b16528191633eb39eef4002d49ace7ddb0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor@sha256:876a222b97b38b35012883c4146c8d102d019fcbe79f26d731d6f2e225e22ffc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:ff0c553ceeb2e0f44b010e37dc6d0db8a251797b88e56468b7cf7f05253e4232,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:624f553f073af7493d34828b074adc9981cce403edd8e71482c7307008479fd9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:e3874936a518c8560339db8f840fc5461885819f6050b5de8d3ab9199bea5094,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:1cea25f1d2a45affc80c46fb9d427749d3f06b61590ac6070a2910e3ec8a4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:e36d5b9a65194f12f7b01c6422ba3ed52a687fd1695fbb21f4986c67d9f9317f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:8b21bec527d54cd766e277889df6bcccd2baeaa946274606b986c0c3b7ca689f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:45aceca77f8fcf61127f0da650bdfdf11ede9b0944c78b63fab819d03283f96b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:709ac58998927dd61786821ae1e63343fd97ccf5763aac5edb4583eea9401d22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:867d4ef7c21f75e6030a685b5762ab4d84b671316ed6b98d75200076e93342cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2b90da93550b99d2fcfa95bd819f3363aa68346a416f8dc7baac3e9c5f487761,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:6f86db36d668348be8c5b46dcda8b1fa23d34bfdc07164fbcbe7a6327fb4de24,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:8cde52cef8795d1c91983b100d86541c7718160ec260fe0f97b96add4c2c8ee8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:a9583cb3baf440d2358ef041373833afbeae60da8159dd031502379901141620,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:835ebed082fe1c45bd799d1d5357595ce63efeb05ca876f26b08443facb9c164,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:011d682241db724bc40736c9b54d2ea450ea7e6be095b1ff5fa28c8007466775,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:2025da90cff8f563deb08bee71efe16d4078edc2a767b2e225cca5c77f1aa2f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:ff46cd5e0e13d105c4629e78c2734a50835f06b6a1e31da9e0462981d10c4be3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:5b4fd0c2b76fa5539f74687b11c5882d77bd31352452322b37ff51fa18f12a61,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:5f6045841aff0fde6f684a34cdf49f8dc7b2c3bcbdeab201f1058971e0c5f79e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:448f4e1b740c30936e340bd6e8534d78c83357bf373a4223950aa64d3484f007,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:b68e3615af8a0eb0ef6bf9ceeef59540a6f4a9a85f6078a3620be115c73a7db8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:7eae01cf60383e523c9cd94d158a9162120a7370829a1dad20fdea6b0fd660bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:28cc10501788081eb61b5a1af35546191a92741f4f109df54c74e2b19439d0f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:9a616e37acfd120612f78043237a8541266ba34883833c9beb43f3da313661ad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:6b1be6cd94a0942259bca5d5d2c30cc7de4a33276b61f8ae3940226772106256,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:02d2c22d15401574941fbe057095442dee0d6f7a0a9341de35d25e6a12a3fe4b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:fc3b3a36b74fd653946723c54b208072d52200635850b531e9d595a7aaea5a01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:7850ccbff320bf9a1c9c769c1c70777eb97117dd8cd5ae4435be9b4622cf807a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:397dac7e39cf40d14a986e6ec4a60fb698ca35c197d0db315b1318514cc6d1d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:1c95142a36276686e720f86423ee171dc9adcc1e89879f627545b7c906ccd9bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:e331a8fde6638e5ba154c4f0b38772a9a424f60656f2777245975fb1fa02f07d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:b6e1e8a249d36ef36c6ac4170af1e043dda1ccc0f9672832d3ff151bf3533076,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:cd3cf7a34053e850b4d4f9f4ea4c74953a54a42fd18e47d7c01d44a88923e925,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:aee28476344fc0cc148fbe97daf9b1bfcedc22001550bba4bdc4e84be7b6989d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:cfa0b92c976603ee2a937d34013a238fcd8aa75f998e50642e33489f14124633,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:73c2f2d6eecf88acf4e45b133c8373d9bb006b530e0aff0b28f3b7420620a874,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:927b405cc04abe5ff716186e8d35e2dc5fad1c8430194659ee6617d74e4e055d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:6154d7cebd7c339afa5b86330262156171743aa5b79c2b78f9a2f378005ed8fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:e2db2f4af8d3d0be7868c6efef0189f3a2c74a8f96ae10e3f991cdf83feaef29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:c773629df257726a6d3cacc24a6e4df0babcd7d37df04e6d14676a8da028b9c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:776211111e2e6493706dbc49a3ba44f31d1b947919313ed3a0f35810e304ec52,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:ae4a20d9aad04cfaeaa3105fa8e37db4216c3b17530bc98daf1204555bc23485,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:7cccf24ad0a152f90ca39893064f48a1656950ee8142685a5d482c71f0bdc9f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:af46761060c7987e1dee5f14c06d85b46f12ad8e09c83d4246ab4e3a65dfda3e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:05450b48f6b5352b2686a26e933e8727748edae2ae9652d9164b7d7a1817c55a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:fc9c99eeef91523482bd8f92661b393287e1f2a24ad2ba9e33191f8de9af74cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:3e4ecc02b4b5e0860482a93599ba9ca598c5ce26c093c46e701f96fe51acb208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:2346037e064861c7892690d2e8b3e1eea1a26ce3c3a11fda0b41301965bc828c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:c26c3ff9cabe3593ceb10006e782bf9391ac14785768ce9eec4f938c2d3cf228,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:daa45220bb1c47922d0917aa8fe423bb82b03a01429f1c9e37635e701e352d71,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:58ac66ca1be01fe0157977bd79a26cde4d0de153edfaf4162367c924826b2ef4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:5e3f93f3085cfd94e599bbf771635477e5e015b7c22c624edca926459d369e69,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:6edd7f91b0fc53dd91194f6e0c206a98e5667bb7a9c5f2a423349612d7300506,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:2590b6c6197091ca423dfb93a609e0d843b270ad642f0c1920ac23f79aec8dca,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pg6vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-b58f89467-gqwwm_openstack-operators(b10f4358-81d9-4d88-a0c5-3aa698f00cdf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:30:15 crc kubenswrapper[5035]: I1124 19:30:15.234025 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:30:15 crc kubenswrapper[5035]: I1124 19:30:15.234082 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:30:15 crc kubenswrapper[5035]: I1124 19:30:15.968473 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" event={"ID":"95f16522-f695-414f-8ec6-c2cb86913080","Type":"ContainerStarted","Data":"10016d93464ed15cc1ef58d4a713426b78131ab0b97078e79b3dd7e72d40ac0d"} Nov 24 19:30:15 crc kubenswrapper[5035]: I1124 19:30:15.974018 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" event={"ID":"d99686fc-e38b-49de-89a3-7de31352efc3","Type":"ContainerStarted","Data":"2392db77ace91ed23880f39dc46b41c056be265167b36c972a2015f695d15f72"} Nov 24 19:30:15 crc kubenswrapper[5035]: I1124 19:30:15.999579 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" event={"ID":"f4b3bf02-333f-46c2-90a0-fe92b166328c","Type":"ContainerStarted","Data":"b0408f9e281f9e8274a2cf8a89612f849811be9c83c42cde028bb789adb069f1"} Nov 24 19:30:16 crc kubenswrapper[5035]: I1124 19:30:16.000652 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4"] Nov 24 19:30:16 crc kubenswrapper[5035]: I1124 19:30:16.011637 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" event={"ID":"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f","Type":"ContainerStarted","Data":"b6d785fd0e3c7fa13bfe2657edeaf7e859f071e95435a54c8607095f058ff7d8"} Nov 24 19:30:16 crc kubenswrapper[5035]: W1124 19:30:16.052338 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4eb72ce7_5f41_4360_91be_e7bce1db7b69.slice/crio-c09950393ebf67ec218df823b940c91ee73beece060589e95653ba1b79870cd1 WatchSource:0}: Error finding container c09950393ebf67ec218df823b940c91ee73beece060589e95653ba1b79870cd1: Status 404 returned error can't find the container with id c09950393ebf67ec218df823b940c91ee73beece060589e95653ba1b79870cd1 Nov 24 19:30:16 crc kubenswrapper[5035]: E1124 19:30:16.180753 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tzlgv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-748dc6576f-tfjff_openstack-operators(af7760bb-5727-4a40-b9a8-9c7b48a9d066): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 19:30:16 crc kubenswrapper[5035]: E1124 19:30:16.181899 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" podUID="af7760bb-5727-4a40-b9a8-9c7b48a9d066" Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.022371 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" event={"ID":"6db4802c-aa26-4fc1-baa7-37ac2845a460","Type":"ContainerStarted","Data":"2c4649d8d601f2ba4d0e21f15d9dee2651fcabb530225bffc64f75cdb0c22c27"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.023992 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" event={"ID":"bf589985-cf9b-4133-8d16-794c8fbfa0f3","Type":"ContainerStarted","Data":"3d71d4f3b18e827008ccc7503be456de390e4e0bd5c6ca4af0107d297538b917"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.025661 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" event={"ID":"6084f8a1-4cde-4714-af8d-76f274dd4b68","Type":"ContainerStarted","Data":"16b613230634590ccdbb84e21612798990c34f34a5ec507009492179d7a8a02f"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.026922 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" event={"ID":"0f776395-f2dd-41cc-af5f-e38dcd9da7b3","Type":"ContainerStarted","Data":"b34c1c5a433143855762bc4a3442ee19758af7188d7b4a72d35170a25005cbb8"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.028475 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" event={"ID":"f835e1e6-2181-4243-9572-0566b72f2f11","Type":"ContainerStarted","Data":"ab58d0bee5b3b9078cf7723c95b374563775b4cdae44fb9b7780f58498073c70"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.029979 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" event={"ID":"af7760bb-5727-4a40-b9a8-9c7b48a9d066","Type":"ContainerStarted","Data":"6820061c9c192030ce3a420709170e32b06684c35de911d79a9f82b979d06504"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.030109 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.031092 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" event={"ID":"792f458f-945a-4846-ac66-c3e2801c2be6","Type":"ContainerStarted","Data":"37f8ab6351a43fe1d24c7faa920eefada7d686a8cde5435fa06dded0d2ed07a2"} Nov 24 19:30:17 crc kubenswrapper[5035]: E1124 19:30:17.031275 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" podUID="af7760bb-5727-4a40-b9a8-9c7b48a9d066" Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.033749 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" event={"ID":"4eb72ce7-5f41-4360-91be-e7bce1db7b69","Type":"ContainerStarted","Data":"9e9b34ccbb8f28cc2d4cd4d55667be3222ecae5a7c17e89f318f91bcb149c06d"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.033787 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" event={"ID":"4eb72ce7-5f41-4360-91be-e7bce1db7b69","Type":"ContainerStarted","Data":"c09950393ebf67ec218df823b940c91ee73beece060589e95653ba1b79870cd1"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.034172 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.035204 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" event={"ID":"f779482e-56ae-49de-8d6d-e06bf57dc3a7","Type":"ContainerStarted","Data":"1a8c7104cea986a0d8e2c4f2ae2a7c6bde3c7cd5030349d958ff1458770b82db"} Nov 24 19:30:17 crc kubenswrapper[5035]: I1124 19:30:17.090552 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" podStartSLOduration=16.090533091 podStartE2EDuration="16.090533091s" podCreationTimestamp="2025-11-24 19:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:30:17.083593991 +0000 UTC m=+955.606100258" watchObservedRunningTime="2025-11-24 19:30:17.090533091 +0000 UTC m=+955.613039348" Nov 24 19:30:18 crc kubenswrapper[5035]: E1124 19:30:18.045571 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" podUID="af7760bb-5727-4a40-b9a8-9c7b48a9d066" Nov 24 19:30:21 crc kubenswrapper[5035]: I1124 19:30:21.268575 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" Nov 24 19:30:21 crc kubenswrapper[5035]: E1124 19:30:21.276623 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" podUID="af7760bb-5727-4a40-b9a8-9c7b48a9d066" Nov 24 19:30:25 crc kubenswrapper[5035]: I1124 19:30:25.269538 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-hvjd4" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.159934 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.160562 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rmkf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-79856dc55c-wp2d7_openstack-operators(95f16522-f695-414f-8ec6-c2cb86913080): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.162536 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" podUID="95f16522-f695-414f-8ec6-c2cb86913080" Nov 24 19:30:34 crc kubenswrapper[5035]: I1124 19:30:34.174463 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" Nov 24 19:30:34 crc kubenswrapper[5035]: I1124 19:30:34.177963 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.220923 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" podUID="95f16522-f695-414f-8ec6-c2cb86913080" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.411003 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.411201 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6tb79,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7c57c8bbc4-5wtlz_openstack-operators(6db4802c-aa26-4fc1-baa7-37ac2845a460): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.412740 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" podUID="6db4802c-aa26-4fc1-baa7-37ac2845a460" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.412769 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.412997 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gfvn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-68b95954c9-ws7tg_openstack-operators(d99686fc-e38b-49de-89a3-7de31352efc3): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Nov 24 19:30:34 crc kubenswrapper[5035]: E1124 19:30:34.414727 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" podUID="d99686fc-e38b-49de-89a3-7de31352efc3" Nov 24 19:30:35 crc kubenswrapper[5035]: I1124 19:30:35.182264 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" Nov 24 19:30:35 crc kubenswrapper[5035]: I1124 19:30:35.182586 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.187416 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" podUID="6db4802c-aa26-4fc1-baa7-37ac2845a460" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.187487 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" podUID="d99686fc-e38b-49de-89a3-7de31352efc3" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.189513 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" podUID="95f16522-f695-414f-8ec6-c2cb86913080" Nov 24 19:30:35 crc kubenswrapper[5035]: I1124 19:30:35.189954 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" Nov 24 19:30:35 crc kubenswrapper[5035]: I1124 19:30:35.190031 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.203936 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.204276 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-22w99,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-774b86978c-ckj7f_openstack-operators(9ae5fef4-4b29-4ea2-8404-b145d960ef4a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.205886 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" podUID="9ae5fef4-4b29-4ea2-8404-b145d960ef4a" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.229457 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.229603 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pg6vv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-b58f89467-gqwwm_openstack-operators(b10f4358-81d9-4d88-a0c5-3aa698f00cdf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:30:35 crc kubenswrapper[5035]: E1124 19:30:35.230879 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" podUID="b10f4358-81d9-4d88-a0c5-3aa698f00cdf" Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.264808 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" event={"ID":"be99a698-bf30-4dd1-b534-e785a37143bc","Type":"ContainerStarted","Data":"22b7a500b7c02f70df5557b3c7582b9e91bd3841ec4268a4a33f2489f804d56b"} Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.265338 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" event={"ID":"00919c94-bf31-4ea8-a322-5b81bdb050c9","Type":"ContainerStarted","Data":"49389a8c75331588d9d03c098cce4763559ab4b84824046f73cdcdc962c9a842"} Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.273626 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" event={"ID":"5d0cb726-bb99-44fe-97c5-000e9619a714","Type":"ContainerStarted","Data":"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a"} Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.286917 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" event={"ID":"98c2e05e-f4af-4979-960b-4104a843321a","Type":"ContainerStarted","Data":"c5909fbe2ecaf34a2c4bf733bff0702d939cfe07a8ab54365ebf8c577a7c9770"} Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.307811 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-n2s8q" podStartSLOduration=3.527243516 podStartE2EDuration="35.30779635s" podCreationTimestamp="2025-11-24 19:30:01 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.809373741 +0000 UTC m=+941.331879998" lastFinishedPulling="2025-11-24 19:30:34.589926555 +0000 UTC m=+973.112432832" observedRunningTime="2025-11-24 19:30:36.306268689 +0000 UTC m=+974.828774936" watchObservedRunningTime="2025-11-24 19:30:36.30779635 +0000 UTC m=+974.830302597" Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.311751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" event={"ID":"eab64b25-1263-4d64-ba86-1f7227feae79","Type":"ContainerStarted","Data":"a8aebb8ab597ae7660c4cf183becba5a653e20edab47196c7b5528c3a56303b6"} Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.319734 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" event={"ID":"ba2f4f9d-09b7-4192-9061-39d5204dd054","Type":"ContainerStarted","Data":"650dd237f42b36aed349e823a6ff7ffc5ce60798f48eebe5ba7d22faeab78fd6"} Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.326448 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" event={"ID":"8ba349a2-1cc6-43ef-9045-4d02feffffd9","Type":"ContainerStarted","Data":"a74981a76435ee517a7b92996ab2556f14ef9b1e0b8906db1640f86cbc35d8ad"} Nov 24 19:30:36 crc kubenswrapper[5035]: I1124 19:30:36.329425 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" event={"ID":"3d6e2a89-1911-4c26-8a9c-03c9955af913","Type":"ContainerStarted","Data":"cf4b52fcf518c9e69ed5237de1abcac600fc1f0ddc46db59da011b00337ca7c3"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.359845 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" event={"ID":"af7760bb-5727-4a40-b9a8-9c7b48a9d066","Type":"ContainerStarted","Data":"5dd0d8f12addd0386d0f417bd1f2e2bb78fac944a530a5e4ed3e95037563d3b9"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.370888 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" event={"ID":"792f458f-945a-4846-ac66-c3e2801c2be6","Type":"ContainerStarted","Data":"5c16863c0dec874040dda1d2837fd373e3d6e6bdfaeb93641bb62ebe20ffc51d"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.371205 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.372882 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.380256 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" event={"ID":"be99a698-bf30-4dd1-b534-e785a37143bc","Type":"ContainerStarted","Data":"0c32070ea2ebde7ebb19d35f01047c2263d337d90575931a54c88273137f89d9"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.380402 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.383489 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" event={"ID":"00919c94-bf31-4ea8-a322-5b81bdb050c9","Type":"ContainerStarted","Data":"a83495a366013a8801a462f1874990cc391ebb62423d7be1ffd3384a94a9e9f5"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.383615 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.386961 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-tfjff" podStartSLOduration=24.445976162 podStartE2EDuration="37.386940936s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.5661411 +0000 UTC m=+941.088647357" lastFinishedPulling="2025-11-24 19:30:15.507105854 +0000 UTC m=+954.029612131" observedRunningTime="2025-11-24 19:30:37.384352155 +0000 UTC m=+975.906858412" watchObservedRunningTime="2025-11-24 19:30:37.386940936 +0000 UTC m=+975.909447193" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.388596 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" event={"ID":"bf589985-cf9b-4133-8d16-794c8fbfa0f3","Type":"ContainerStarted","Data":"386b9646b9d7d8862d440507f34b94477b0dec1d7fdb3d20ffa2b49c91b5f013"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.388726 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.390900 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.403776 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" event={"ID":"5d0cb726-bb99-44fe-97c5-000e9619a714","Type":"ContainerStarted","Data":"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.403907 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.409625 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" event={"ID":"f779482e-56ae-49de-8d6d-e06bf57dc3a7","Type":"ContainerStarted","Data":"0e1c6b622fcb39c6743cb6c4ada278f92dc97b54745bad2745f813f2854bcefe"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.409853 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.412076 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.416506 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" event={"ID":"d99686fc-e38b-49de-89a3-7de31352efc3","Type":"ContainerStarted","Data":"e0dfee3eb7a6080bd26c8342bb5d9e998a831a92f48b64b5387950cb16d235a6"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.423777 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" podStartSLOduration=4.994852978 podStartE2EDuration="37.423762494s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.8170021 +0000 UTC m=+941.339508357" lastFinishedPulling="2025-11-24 19:30:35.245911616 +0000 UTC m=+973.768417873" observedRunningTime="2025-11-24 19:30:37.411102717 +0000 UTC m=+975.933608974" watchObservedRunningTime="2025-11-24 19:30:37.423762494 +0000 UTC m=+975.946268751" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.444864 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" event={"ID":"9ae5fef4-4b29-4ea2-8404-b145d960ef4a","Type":"ContainerStarted","Data":"ddeda1c231413e235fec3be54926583c075a2201ec937dfb147f6958b49c0438"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.462642 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-csfwz" podStartSLOduration=3.3404061609999998 podStartE2EDuration="37.462621308s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.548269749 +0000 UTC m=+941.070776006" lastFinishedPulling="2025-11-24 19:30:36.670484896 +0000 UTC m=+975.192991153" observedRunningTime="2025-11-24 19:30:37.445115218 +0000 UTC m=+975.967621475" watchObservedRunningTime="2025-11-24 19:30:37.462621308 +0000 UTC m=+975.985127565" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.467586 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" event={"ID":"b10f4358-81d9-4d88-a0c5-3aa698f00cdf","Type":"ContainerStarted","Data":"5577d92662983a4895476f78119b3ad287d7ec1c9274ae21228102f1fc281b27"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.503936 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" podStartSLOduration=6.924310742 podStartE2EDuration="37.503920169s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.80240832 +0000 UTC m=+941.324914577" lastFinishedPulling="2025-11-24 19:30:33.382017747 +0000 UTC m=+971.904524004" observedRunningTime="2025-11-24 19:30:37.474667447 +0000 UTC m=+975.997173704" watchObservedRunningTime="2025-11-24 19:30:37.503920169 +0000 UTC m=+976.026426426" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.506185 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-ws7tg" podStartSLOduration=24.244730737 podStartE2EDuration="37.506176901s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.240269782 +0000 UTC m=+940.762776039" lastFinishedPulling="2025-11-24 19:30:15.501715936 +0000 UTC m=+954.024222203" observedRunningTime="2025-11-24 19:30:37.501886993 +0000 UTC m=+976.024393250" watchObservedRunningTime="2025-11-24 19:30:37.506176901 +0000 UTC m=+976.028683158" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.507826 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" event={"ID":"3d6e2a89-1911-4c26-8a9c-03c9955af913","Type":"ContainerStarted","Data":"c6ac2efdf411aa66de4ec67df1a3cc2d7b70dc091bea6f76f066db000a4a7d0a"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.507898 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.532541 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" event={"ID":"0f776395-f2dd-41cc-af5f-e38dcd9da7b3","Type":"ContainerStarted","Data":"7b23f871b1170b4d2b8672a2bf382fe5849f900a4703bbd2047f672630a5d5af"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.533415 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.539545 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.547930 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" podStartSLOduration=5.205757663 podStartE2EDuration="36.547914774s" podCreationTimestamp="2025-11-24 19:30:01 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.81701548 +0000 UTC m=+941.339521727" lastFinishedPulling="2025-11-24 19:30:34.159172571 +0000 UTC m=+972.681678838" observedRunningTime="2025-11-24 19:30:37.532037538 +0000 UTC m=+976.054543795" watchObservedRunningTime="2025-11-24 19:30:37.547914774 +0000 UTC m=+976.070421031" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.563270 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" event={"ID":"f4b3bf02-333f-46c2-90a0-fe92b166328c","Type":"ContainerStarted","Data":"12ff7801a6ec1da081a6e7de3bf97c3ad1a31c9b98870b9ac808dac9d37372c7"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.564013 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-6rjkn" podStartSLOduration=3.662340955 podStartE2EDuration="37.563997834s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:01.853605627 +0000 UTC m=+940.376111884" lastFinishedPulling="2025-11-24 19:30:35.755262506 +0000 UTC m=+974.277768763" observedRunningTime="2025-11-24 19:30:37.561802193 +0000 UTC m=+976.084308450" watchObservedRunningTime="2025-11-24 19:30:37.563997834 +0000 UTC m=+976.086504091" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.564132 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.569178 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.575863 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" event={"ID":"c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f","Type":"ContainerStarted","Data":"40b0f4517c1bac90ac9fe85037d52efb7c773af147348a8e217e4057764ea739"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.576827 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.588037 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-kkcw8" podStartSLOduration=4.60948618 podStartE2EDuration="37.588022712s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.573508371 +0000 UTC m=+941.096014628" lastFinishedPulling="2025-11-24 19:30:35.552044903 +0000 UTC m=+974.074551160" observedRunningTime="2025-11-24 19:30:37.583730694 +0000 UTC m=+976.106236951" watchObservedRunningTime="2025-11-24 19:30:37.588022712 +0000 UTC m=+976.110528969" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.593747 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.608875 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" event={"ID":"6db4802c-aa26-4fc1-baa7-37ac2845a460","Type":"ContainerStarted","Data":"1782f64f21780093265f360567968078d00162aeb4582fd02ddc75effb7186fe"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.616220 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-p8vr6" podStartSLOduration=4.116037806 podStartE2EDuration="37.616204123s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.260449196 +0000 UTC m=+940.782955453" lastFinishedPulling="2025-11-24 19:30:35.760615513 +0000 UTC m=+974.283121770" observedRunningTime="2025-11-24 19:30:37.615254398 +0000 UTC m=+976.137760655" watchObservedRunningTime="2025-11-24 19:30:37.616204123 +0000 UTC m=+976.138710380" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.619217 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" event={"ID":"8ba349a2-1cc6-43ef-9045-4d02feffffd9","Type":"ContainerStarted","Data":"054be28c5fd62dc98e8a75e5fd1b33f72c0c2a1bec02029d58e483c4776686c8"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.619491 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.641615 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" event={"ID":"6084f8a1-4cde-4714-af8d-76f274dd4b68","Type":"ContainerStarted","Data":"0adf2677b1c45d1e8532191b01c64056efa2b664fd4861d96c201e62e6d86738"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.642650 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.645624 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-bvfmn" podStartSLOduration=5.125683636 podStartE2EDuration="37.645602118s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.790984427 +0000 UTC m=+941.313490684" lastFinishedPulling="2025-11-24 19:30:35.310902909 +0000 UTC m=+973.833409166" observedRunningTime="2025-11-24 19:30:37.641654151 +0000 UTC m=+976.164160408" watchObservedRunningTime="2025-11-24 19:30:37.645602118 +0000 UTC m=+976.168108375" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.646241 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.656726 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" event={"ID":"f835e1e6-2181-4243-9572-0566b72f2f11","Type":"ContainerStarted","Data":"f2c26c992471b5ca2d806334e43d80cd5af48444d13b2c5239594e84bf976754"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.657706 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.666945 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-mq8nd" podStartSLOduration=4.680178266 podStartE2EDuration="37.666925403s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.566061297 +0000 UTC m=+941.088567634" lastFinishedPulling="2025-11-24 19:30:35.552808514 +0000 UTC m=+974.075314771" observedRunningTime="2025-11-24 19:30:37.663750135 +0000 UTC m=+976.186256402" watchObservedRunningTime="2025-11-24 19:30:37.666925403 +0000 UTC m=+976.189431670" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.668569 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.675555 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" event={"ID":"eab64b25-1263-4d64-ba86-1f7227feae79","Type":"ContainerStarted","Data":"9eacd76866b0d629ebec01c7d98bf6ce08746be4eeb2af6f51424333902e55eb"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.676316 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.690573 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" event={"ID":"ba2f4f9d-09b7-4192-9061-39d5204dd054","Type":"ContainerStarted","Data":"d904492d1512933d885089a9b971ee5a99b7358a9e261a67b14fec216485005c"} Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.691351 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.733961 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" podStartSLOduration=4.935506831 podStartE2EDuration="36.733934907s" podCreationTimestamp="2025-11-24 19:30:01 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.791577872 +0000 UTC m=+941.314084129" lastFinishedPulling="2025-11-24 19:30:34.590005948 +0000 UTC m=+973.112512205" observedRunningTime="2025-11-24 19:30:37.691583898 +0000 UTC m=+976.214090155" watchObservedRunningTime="2025-11-24 19:30:37.733934907 +0000 UTC m=+976.256441174" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.734162 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" podStartSLOduration=8.321866522 podStartE2EDuration="37.734156184s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.802359648 +0000 UTC m=+941.324865905" lastFinishedPulling="2025-11-24 19:30:32.21464931 +0000 UTC m=+970.737155567" observedRunningTime="2025-11-24 19:30:37.728705325 +0000 UTC m=+976.251211592" watchObservedRunningTime="2025-11-24 19:30:37.734156184 +0000 UTC m=+976.256662441" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.759825 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" podStartSLOduration=5.983163858 podStartE2EDuration="37.759806136s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.815816257 +0000 UTC m=+941.338322514" lastFinishedPulling="2025-11-24 19:30:34.592458515 +0000 UTC m=+973.114964792" observedRunningTime="2025-11-24 19:30:37.755601961 +0000 UTC m=+976.278108218" watchObservedRunningTime="2025-11-24 19:30:37.759806136 +0000 UTC m=+976.282312393" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.842602 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-ln7hm" podStartSLOduration=4.43300898 podStartE2EDuration="37.842586453s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.345045866 +0000 UTC m=+940.867552123" lastFinishedPulling="2025-11-24 19:30:35.754623339 +0000 UTC m=+974.277129596" observedRunningTime="2025-11-24 19:30:37.840716572 +0000 UTC m=+976.363222839" watchObservedRunningTime="2025-11-24 19:30:37.842586453 +0000 UTC m=+976.365092710" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.924928 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" podStartSLOduration=6.127926321 podStartE2EDuration="37.924902217s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.792845837 +0000 UTC m=+941.315352094" lastFinishedPulling="2025-11-24 19:30:34.589821733 +0000 UTC m=+973.112327990" observedRunningTime="2025-11-24 19:30:37.905854635 +0000 UTC m=+976.428360892" watchObservedRunningTime="2025-11-24 19:30:37.924902217 +0000 UTC m=+976.447408484" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.930182 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-rd2qm" podStartSLOduration=4.505836955 podStartE2EDuration="37.930162382s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.326023765 +0000 UTC m=+940.848530022" lastFinishedPulling="2025-11-24 19:30:35.750349192 +0000 UTC m=+974.272855449" observedRunningTime="2025-11-24 19:30:37.869567862 +0000 UTC m=+976.392074119" watchObservedRunningTime="2025-11-24 19:30:37.930162382 +0000 UTC m=+976.452668639" Nov 24 19:30:37 crc kubenswrapper[5035]: I1124 19:30:37.934467 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-5wtlz" podStartSLOduration=24.999346926 podStartE2EDuration="37.934448299s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.566023466 +0000 UTC m=+941.088529723" lastFinishedPulling="2025-11-24 19:30:15.501124839 +0000 UTC m=+954.023631096" observedRunningTime="2025-11-24 19:30:37.929883223 +0000 UTC m=+976.452389480" watchObservedRunningTime="2025-11-24 19:30:37.934448299 +0000 UTC m=+976.456954556" Nov 24 19:30:38 crc kubenswrapper[5035]: I1124 19:30:38.702813 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" event={"ID":"9ae5fef4-4b29-4ea2-8404-b145d960ef4a","Type":"ContainerStarted","Data":"17d6efa34e9b4fd374369b300c02ecd08ac4e1e44e15a53fcf56ffda7ab8dfc4"} Nov 24 19:30:38 crc kubenswrapper[5035]: I1124 19:30:38.703019 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" Nov 24 19:30:38 crc kubenswrapper[5035]: I1124 19:30:38.705819 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" event={"ID":"b10f4358-81d9-4d88-a0c5-3aa698f00cdf","Type":"ContainerStarted","Data":"f6d019c1cfee47269ca92a6fe5864963f1b381ad5057aa0ad4ce827ae70ea6cb"} Nov 24 19:30:38 crc kubenswrapper[5035]: I1124 19:30:38.729567 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" podStartSLOduration=4.272271841 podStartE2EDuration="38.729543742s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.548576818 +0000 UTC m=+941.071083075" lastFinishedPulling="2025-11-24 19:30:37.005848719 +0000 UTC m=+975.528354976" observedRunningTime="2025-11-24 19:30:38.72327007 +0000 UTC m=+977.245776337" watchObservedRunningTime="2025-11-24 19:30:38.729543742 +0000 UTC m=+977.252050029" Nov 24 19:30:38 crc kubenswrapper[5035]: I1124 19:30:38.758043 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" podStartSLOduration=5.470865353 podStartE2EDuration="38.758011432s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:03.716975593 +0000 UTC m=+942.239481860" lastFinishedPulling="2025-11-24 19:30:37.004121682 +0000 UTC m=+975.526627939" observedRunningTime="2025-11-24 19:30:38.754915477 +0000 UTC m=+977.277421734" watchObservedRunningTime="2025-11-24 19:30:38.758011432 +0000 UTC m=+977.280517699" Nov 24 19:30:39 crc kubenswrapper[5035]: I1124 19:30:39.715420 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:41 crc kubenswrapper[5035]: I1124 19:30:41.471154 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rqf69" Nov 24 19:30:41 crc kubenswrapper[5035]: I1124 19:30:41.546527 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-prr8k" Nov 24 19:30:41 crc kubenswrapper[5035]: I1124 19:30:41.577699 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-87rk8" Nov 24 19:30:41 crc kubenswrapper[5035]: I1124 19:30:41.604602 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-tpw7d" Nov 24 19:30:41 crc kubenswrapper[5035]: I1124 19:30:41.623939 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 19:30:41 crc kubenswrapper[5035]: I1124 19:30:41.637005 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-rjs7d" Nov 24 19:30:41 crc kubenswrapper[5035]: I1124 19:30:41.661864 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-grqcw" Nov 24 19:30:43 crc kubenswrapper[5035]: I1124 19:30:43.111150 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gqwwm" Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.234498 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.234890 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.234946 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.235610 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f8c76c75a096d1e632ca8dbc062a2c86c922da2cd856db7b41a1460c00d789d"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.235674 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://7f8c76c75a096d1e632ca8dbc062a2c86c922da2cd856db7b41a1460c00d789d" gracePeriod=600 Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.757797 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="7f8c76c75a096d1e632ca8dbc062a2c86c922da2cd856db7b41a1460c00d789d" exitCode=0 Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.757837 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"7f8c76c75a096d1e632ca8dbc062a2c86c922da2cd856db7b41a1460c00d789d"} Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.757861 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"d2a1e9879f1117f6209b6479ae153039678ddcd07bf4e0e5400eeea2f3c473b0"} Nov 24 19:30:45 crc kubenswrapper[5035]: I1124 19:30:45.757877 5035 scope.go:117] "RemoveContainer" containerID="88531e6ce05adbac2436f424c2cf6aa59384e9fc263400422a229df68ffdfb39" Nov 24 19:30:49 crc kubenswrapper[5035]: I1124 19:30:49.791351 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" event={"ID":"95f16522-f695-414f-8ec6-c2cb86913080","Type":"ContainerStarted","Data":"bfec4101f457245c46c2c9baba8d8428ec81e1b1dcb5187e0b4559a2e9c1bfeb"} Nov 24 19:30:49 crc kubenswrapper[5035]: I1124 19:30:49.839612 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-wp2d7" podStartSLOduration=36.650664099 podStartE2EDuration="49.839587893s" podCreationTimestamp="2025-11-24 19:30:00 +0000 UTC" firstStartedPulling="2025-11-24 19:30:02.312201725 +0000 UTC m=+940.834707982" lastFinishedPulling="2025-11-24 19:30:15.501125519 +0000 UTC m=+954.023631776" observedRunningTime="2025-11-24 19:30:49.810828566 +0000 UTC m=+988.333334833" watchObservedRunningTime="2025-11-24 19:30:49.839587893 +0000 UTC m=+988.362094160" Nov 24 19:30:51 crc kubenswrapper[5035]: I1124 19:30:51.123941 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-ckj7f" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.751721 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-ndp97"] Nov 24 19:31:04 crc kubenswrapper[5035]: E1124 19:31:04.752678 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e46e48-d877-4332-918a-95d8c6afc3b5" containerName="collect-profiles" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.752694 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e46e48-d877-4332-918a-95d8c6afc3b5" containerName="collect-profiles" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.752853 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e46e48-d877-4332-918a-95d8c6afc3b5" containerName="collect-profiles" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.753774 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.757121 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5m76c" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.757176 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.757225 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.757181 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.757709 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-ndp97"] Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.810045 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6584b49599-bwjn6"] Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.812096 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.813998 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.816541 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-bwjn6"] Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.862955 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab18ea7-92bb-4693-a40c-148f07c2862f-config\") pod \"dnsmasq-dns-7bdd77c89-ndp97\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.863033 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzdmt\" (UniqueName: \"kubernetes.io/projected/5ab18ea7-92bb-4693-a40c-148f07c2862f-kube-api-access-pzdmt\") pod \"dnsmasq-dns-7bdd77c89-ndp97\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.964449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzdmt\" (UniqueName: \"kubernetes.io/projected/5ab18ea7-92bb-4693-a40c-148f07c2862f-kube-api-access-pzdmt\") pod \"dnsmasq-dns-7bdd77c89-ndp97\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.964529 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-config\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.964579 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab18ea7-92bb-4693-a40c-148f07c2862f-config\") pod \"dnsmasq-dns-7bdd77c89-ndp97\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.964605 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqxdz\" (UniqueName: \"kubernetes.io/projected/b56b4d33-ed61-434e-87bb-e0ee1e53513a-kube-api-access-wqxdz\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.964630 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-dns-svc\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.965586 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab18ea7-92bb-4693-a40c-148f07c2862f-config\") pod \"dnsmasq-dns-7bdd77c89-ndp97\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:04 crc kubenswrapper[5035]: I1124 19:31:04.992985 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzdmt\" (UniqueName: \"kubernetes.io/projected/5ab18ea7-92bb-4693-a40c-148f07c2862f-kube-api-access-pzdmt\") pod \"dnsmasq-dns-7bdd77c89-ndp97\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.065442 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-config\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.065764 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqxdz\" (UniqueName: \"kubernetes.io/projected/b56b4d33-ed61-434e-87bb-e0ee1e53513a-kube-api-access-wqxdz\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.065786 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-dns-svc\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.066406 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-config\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.066664 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-dns-svc\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.072980 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.082408 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqxdz\" (UniqueName: \"kubernetes.io/projected/b56b4d33-ed61-434e-87bb-e0ee1e53513a-kube-api-access-wqxdz\") pod \"dnsmasq-dns-6584b49599-bwjn6\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.140130 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.412163 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-bwjn6"] Nov 24 19:31:05 crc kubenswrapper[5035]: W1124 19:31:05.416769 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb56b4d33_ed61_434e_87bb_e0ee1e53513a.slice/crio-5ad9670adc12b83d1e8b591aa349aef515885aad88ad323eb14d6d716fb42f93 WatchSource:0}: Error finding container 5ad9670adc12b83d1e8b591aa349aef515885aad88ad323eb14d6d716fb42f93: Status 404 returned error can't find the container with id 5ad9670adc12b83d1e8b591aa349aef515885aad88ad323eb14d6d716fb42f93 Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.418570 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.536905 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-ndp97"] Nov 24 19:31:05 crc kubenswrapper[5035]: W1124 19:31:05.540708 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ab18ea7_92bb_4693_a40c_148f07c2862f.slice/crio-35132a7dbf2085eaee3b0235f75fcc4c5d58437432395b1b4d2536da5d82d577 WatchSource:0}: Error finding container 35132a7dbf2085eaee3b0235f75fcc4c5d58437432395b1b4d2536da5d82d577: Status 404 returned error can't find the container with id 35132a7dbf2085eaee3b0235f75fcc4c5d58437432395b1b4d2536da5d82d577 Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.909644 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-bwjn6" event={"ID":"b56b4d33-ed61-434e-87bb-e0ee1e53513a","Type":"ContainerStarted","Data":"5ad9670adc12b83d1e8b591aa349aef515885aad88ad323eb14d6d716fb42f93"} Nov 24 19:31:05 crc kubenswrapper[5035]: I1124 19:31:05.911219 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" event={"ID":"5ab18ea7-92bb-4693-a40c-148f07c2862f","Type":"ContainerStarted","Data":"35132a7dbf2085eaee3b0235f75fcc4c5d58437432395b1b4d2536da5d82d577"} Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.011028 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-bwjn6"] Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.025831 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-dqv6q"] Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.029280 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.036368 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-dqv6q"] Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.111648 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-config\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.111748 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmw2j\" (UniqueName: \"kubernetes.io/projected/15c17b6f-1037-43de-8e7a-cf86311cc674-kube-api-access-lmw2j\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.111826 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.213411 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmw2j\" (UniqueName: \"kubernetes.io/projected/15c17b6f-1037-43de-8e7a-cf86311cc674-kube-api-access-lmw2j\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.213711 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.213753 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-config\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.214638 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-config\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.216188 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.242558 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmw2j\" (UniqueName: \"kubernetes.io/projected/15c17b6f-1037-43de-8e7a-cf86311cc674-kube-api-access-lmw2j\") pod \"dnsmasq-dns-7c6d9948dc-dqv6q\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.294926 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-ndp97"] Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.318912 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2cczt"] Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.324159 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.343585 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2cczt"] Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.355087 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.416562 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9xnw\" (UniqueName: \"kubernetes.io/projected/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-kube-api-access-g9xnw\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.416641 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-config\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.416680 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-dns-svc\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.518276 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9xnw\" (UniqueName: \"kubernetes.io/projected/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-kube-api-access-g9xnw\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.519523 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-config\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.526545 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-config\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.527449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-dns-svc\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.527591 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-dns-svc\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.548692 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9xnw\" (UniqueName: \"kubernetes.io/projected/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-kube-api-access-g9xnw\") pod \"dnsmasq-dns-6486446b9f-2cczt\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.655503 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.916412 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-dqv6q"] Nov 24 19:31:08 crc kubenswrapper[5035]: W1124 19:31:08.931860 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15c17b6f_1037_43de_8e7a_cf86311cc674.slice/crio-acaef67ec5e5ead8913323a556b0dc04cece24b3bca342b7059812df061a6369 WatchSource:0}: Error finding container acaef67ec5e5ead8913323a556b0dc04cece24b3bca342b7059812df061a6369: Status 404 returned error can't find the container with id acaef67ec5e5ead8913323a556b0dc04cece24b3bca342b7059812df061a6369 Nov 24 19:31:08 crc kubenswrapper[5035]: I1124 19:31:08.945004 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" event={"ID":"15c17b6f-1037-43de-8e7a-cf86311cc674","Type":"ContainerStarted","Data":"acaef67ec5e5ead8913323a556b0dc04cece24b3bca342b7059812df061a6369"} Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.074812 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2cczt"] Nov 24 19:31:09 crc kubenswrapper[5035]: W1124 19:31:09.080910 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ea95d1a_843d_4a31_8c0d_8f351d47c0bc.slice/crio-259e30c6b7794ed4af98c767a6d24e3555135383f2f9e46eede2d9a35c2d9bd5 WatchSource:0}: Error finding container 259e30c6b7794ed4af98c767a6d24e3555135383f2f9e46eede2d9a35c2d9bd5: Status 404 returned error can't find the container with id 259e30c6b7794ed4af98c767a6d24e3555135383f2f9e46eede2d9a35c2d9bd5 Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.179332 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.181869 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.183818 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.184539 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.185174 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.186086 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.186178 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.186102 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-c9p2l" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.187252 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.194418 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342521 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342554 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342584 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342631 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342717 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342792 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342835 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342909 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-config-data\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342934 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ed0e7cb-164f-4689-8714-1f11bfa25725-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342971 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ed0e7cb-164f-4689-8714-1f11bfa25725-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.342987 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpqlq\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-kube-api-access-mpqlq\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444603 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444672 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-config-data\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444696 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ed0e7cb-164f-4689-8714-1f11bfa25725-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444720 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ed0e7cb-164f-4689-8714-1f11bfa25725-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444738 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpqlq\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-kube-api-access-mpqlq\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444777 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444790 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444809 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444829 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444862 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.444900 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.445360 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.445669 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.445809 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.445883 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.446001 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-config-data\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.446606 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.450714 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ed0e7cb-164f-4689-8714-1f11bfa25725-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.450904 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.450995 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.451890 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ed0e7cb-164f-4689-8714-1f11bfa25725-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.460459 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpqlq\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-kube-api-access-mpqlq\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.474559 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.476071 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.477516 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.482401 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.482803 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.482841 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-t7zgv" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.483889 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.484004 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.484603 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.486893 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.499171 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.513315 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.647682 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.647727 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.647882 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8716e4c9-35b0-4922-afc2-13006d78283a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.647918 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8716e4c9-35b0-4922-afc2-13006d78283a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.647934 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.647970 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.647987 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.648013 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.648033 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.648399 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rb26\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-kube-api-access-7rb26\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.648476 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.750559 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.750666 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.750706 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.750727 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8716e4c9-35b0-4922-afc2-13006d78283a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.750759 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8716e4c9-35b0-4922-afc2-13006d78283a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.751019 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.751052 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.753427 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.753480 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.753508 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.753542 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rb26\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-kube-api-access-7rb26\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.755484 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.759100 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.760102 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.760519 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.760605 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.766708 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8716e4c9-35b0-4922-afc2-13006d78283a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.768928 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.769503 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.769857 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.778115 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8716e4c9-35b0-4922-afc2-13006d78283a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.785409 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rb26\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-kube-api-access-7rb26\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.804207 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.879703 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:09 crc kubenswrapper[5035]: I1124 19:31:09.973659 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" event={"ID":"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc","Type":"ContainerStarted","Data":"259e30c6b7794ed4af98c767a6d24e3555135383f2f9e46eede2d9a35c2d9bd5"} Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:09.998766 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:31:10 crc kubenswrapper[5035]: W1124 19:31:10.027372 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ed0e7cb_164f_4689_8714_1f11bfa25725.slice/crio-3d68d7f492531271729710ef1645bc1020b27d15ad7057c2b794bf8b00746c86 WatchSource:0}: Error finding container 3d68d7f492531271729710ef1645bc1020b27d15ad7057c2b794bf8b00746c86: Status 404 returned error can't find the container with id 3d68d7f492531271729710ef1645bc1020b27d15ad7057c2b794bf8b00746c86 Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.395729 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.972009 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.973883 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.976822 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.977111 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.977228 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gtccm" Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.977398 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.982781 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8716e4c9-35b0-4922-afc2-13006d78283a","Type":"ContainerStarted","Data":"45d8dbb82f0c81105f2938ebaf8dd1ff4ac1a3a394bc0890bc245205a72fb6a6"} Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.983747 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 19:31:10 crc kubenswrapper[5035]: I1124 19:31:10.986501 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.022186 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ed0e7cb-164f-4689-8714-1f11bfa25725","Type":"ContainerStarted","Data":"3d68d7f492531271729710ef1645bc1020b27d15ad7057c2b794bf8b00746c86"} Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079354 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fcbe1807-d08b-4d8a-9c92-954b8da6820d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079711 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-config-data-default\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079745 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-kolla-config\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079789 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079818 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slq7c\" (UniqueName: \"kubernetes.io/projected/fcbe1807-d08b-4d8a-9c92-954b8da6820d-kube-api-access-slq7c\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079854 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbe1807-d08b-4d8a-9c92-954b8da6820d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079915 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.079945 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbe1807-d08b-4d8a-9c92-954b8da6820d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182479 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fcbe1807-d08b-4d8a-9c92-954b8da6820d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182550 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-config-data-default\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182576 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-kolla-config\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182620 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182642 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slq7c\" (UniqueName: \"kubernetes.io/projected/fcbe1807-d08b-4d8a-9c92-954b8da6820d-kube-api-access-slq7c\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182686 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbe1807-d08b-4d8a-9c92-954b8da6820d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182828 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.182884 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbe1807-d08b-4d8a-9c92-954b8da6820d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.183476 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.183685 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fcbe1807-d08b-4d8a-9c92-954b8da6820d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.186786 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-config-data-default\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.193185 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.193652 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbe1807-d08b-4d8a-9c92-954b8da6820d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.193942 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbe1807-d08b-4d8a-9c92-954b8da6820d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.210970 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slq7c\" (UniqueName: \"kubernetes.io/projected/fcbe1807-d08b-4d8a-9c92-954b8da6820d-kube-api-access-slq7c\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.212471 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.213728 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fcbe1807-d08b-4d8a-9c92-954b8da6820d-kolla-config\") pod \"openstack-galera-0\" (UID: \"fcbe1807-d08b-4d8a-9c92-954b8da6820d\") " pod="openstack/openstack-galera-0" Nov 24 19:31:11 crc kubenswrapper[5035]: I1124 19:31:11.350119 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.315017 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.318587 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.324219 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.324225 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.324309 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.324352 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-lkmqk" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.335886 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.424824 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.425122 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.425153 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w7cq\" (UniqueName: \"kubernetes.io/projected/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-kube-api-access-6w7cq\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.425185 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.425208 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.425245 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.425287 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.425365 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526422 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526501 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526540 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526572 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w7cq\" (UniqueName: \"kubernetes.io/projected/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-kube-api-access-6w7cq\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526602 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526625 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526661 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.526702 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.527522 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.527816 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.529132 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.529408 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.529780 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.532338 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.533050 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.533205 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.541574 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.541635 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.541799 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5l6x8" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.549537 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.581046 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.589510 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.592931 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w7cq\" (UniqueName: \"kubernetes.io/projected/62cb1b72-4c4f-4c1e-bf23-0dc8a318560d-kube-api-access-6w7cq\") pod \"openstack-cell1-galera-0\" (UID: \"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d\") " pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.632050 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c207-0fb4-4876-89d1-cb18eadb80af-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.632153 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e665c207-0fb4-4876-89d1-cb18eadb80af-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.632181 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e665c207-0fb4-4876-89d1-cb18eadb80af-kolla-config\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.632213 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e665c207-0fb4-4876-89d1-cb18eadb80af-config-data\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.632258 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clng9\" (UniqueName: \"kubernetes.io/projected/e665c207-0fb4-4876-89d1-cb18eadb80af-kube-api-access-clng9\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.666634 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.735168 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e665c207-0fb4-4876-89d1-cb18eadb80af-config-data\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.735259 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clng9\" (UniqueName: \"kubernetes.io/projected/e665c207-0fb4-4876-89d1-cb18eadb80af-kube-api-access-clng9\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.735336 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c207-0fb4-4876-89d1-cb18eadb80af-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.735380 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e665c207-0fb4-4876-89d1-cb18eadb80af-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.735407 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e665c207-0fb4-4876-89d1-cb18eadb80af-kolla-config\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.736123 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e665c207-0fb4-4876-89d1-cb18eadb80af-config-data\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.736204 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e665c207-0fb4-4876-89d1-cb18eadb80af-kolla-config\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.740817 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e665c207-0fb4-4876-89d1-cb18eadb80af-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.754415 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clng9\" (UniqueName: \"kubernetes.io/projected/e665c207-0fb4-4876-89d1-cb18eadb80af-kube-api-access-clng9\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.761377 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c207-0fb4-4876-89d1-cb18eadb80af-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e665c207-0fb4-4876-89d1-cb18eadb80af\") " pod="openstack/memcached-0" Nov 24 19:31:12 crc kubenswrapper[5035]: I1124 19:31:12.950450 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.366192 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.367706 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.372089 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4xwrh" Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.373023 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.464285 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg62h\" (UniqueName: \"kubernetes.io/projected/cc0b947e-0136-4ccf-8203-21604c41ff7d-kube-api-access-sg62h\") pod \"kube-state-metrics-0\" (UID: \"cc0b947e-0136-4ccf-8203-21604c41ff7d\") " pod="openstack/kube-state-metrics-0" Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.565656 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg62h\" (UniqueName: \"kubernetes.io/projected/cc0b947e-0136-4ccf-8203-21604c41ff7d-kube-api-access-sg62h\") pod \"kube-state-metrics-0\" (UID: \"cc0b947e-0136-4ccf-8203-21604c41ff7d\") " pod="openstack/kube-state-metrics-0" Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.604556 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg62h\" (UniqueName: \"kubernetes.io/projected/cc0b947e-0136-4ccf-8203-21604c41ff7d-kube-api-access-sg62h\") pod \"kube-state-metrics-0\" (UID: \"cc0b947e-0136-4ccf-8203-21604c41ff7d\") " pod="openstack/kube-state-metrics-0" Nov 24 19:31:14 crc kubenswrapper[5035]: I1124 19:31:14.691984 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.337148 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.339154 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.341607 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-fwfrz" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.341823 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.342411 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.342429 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.342405 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.345362 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.435107 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.435531 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.435598 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f6b29d-0844-4946-abaf-d331f8f07ba0-config\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.435672 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2mwm\" (UniqueName: \"kubernetes.io/projected/74f6b29d-0844-4946-abaf-d331f8f07ba0-kube-api-access-w2mwm\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.435863 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74f6b29d-0844-4946-abaf-d331f8f07ba0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.435908 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.436125 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/74f6b29d-0844-4946-abaf-d331f8f07ba0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.436175 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.537993 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.538100 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/74f6b29d-0844-4946-abaf-d331f8f07ba0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.538134 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.538171 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.538189 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.538205 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f6b29d-0844-4946-abaf-d331f8f07ba0-config\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.538226 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2mwm\" (UniqueName: \"kubernetes.io/projected/74f6b29d-0844-4946-abaf-d331f8f07ba0-kube-api-access-w2mwm\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.538281 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74f6b29d-0844-4946-abaf-d331f8f07ba0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.539681 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/74f6b29d-0844-4946-abaf-d331f8f07ba0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.540789 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.540811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/74f6b29d-0844-4946-abaf-d331f8f07ba0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.542279 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74f6b29d-0844-4946-abaf-d331f8f07ba0-config\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.557841 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.565917 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.566185 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2mwm\" (UniqueName: \"kubernetes.io/projected/74f6b29d-0844-4946-abaf-d331f8f07ba0-kube-api-access-w2mwm\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.569123 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/74f6b29d-0844-4946-abaf-d331f8f07ba0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.581580 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"74f6b29d-0844-4946-abaf-d331f8f07ba0\") " pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:18 crc kubenswrapper[5035]: I1124 19:31:18.669868 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.290555 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-82r49"] Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.291762 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.293593 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7g4gv" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.294110 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.294852 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.314015 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6l2z4"] Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.319272 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.328904 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-82r49"] Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.351481 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6l2z4"] Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.450407 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-run\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.450667 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-etc-ovs\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.450780 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-log-ovn\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.450913 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-log\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451012 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-run\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451084 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-run-ovn\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451182 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8st6z\" (UniqueName: \"kubernetes.io/projected/aa0e4fdb-6066-4163-9864-9e52443afdb0-kube-api-access-8st6z\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451307 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89539b2e-806f-4062-85fe-efac9c4469d2-scripts\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451437 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0e4fdb-6066-4163-9864-9e52443afdb0-combined-ca-bundle\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451568 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-lib\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7688c\" (UniqueName: \"kubernetes.io/projected/89539b2e-806f-4062-85fe-efac9c4469d2-kube-api-access-7688c\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451779 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa0e4fdb-6066-4163-9864-9e52443afdb0-scripts\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.451885 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa0e4fdb-6066-4163-9864-9e52443afdb0-ovn-controller-tls-certs\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553248 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-run\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553324 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-etc-ovs\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553356 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-log-ovn\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553379 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-log\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553399 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-run\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553417 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-run-ovn\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553436 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8st6z\" (UniqueName: \"kubernetes.io/projected/aa0e4fdb-6066-4163-9864-9e52443afdb0-kube-api-access-8st6z\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553466 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89539b2e-806f-4062-85fe-efac9c4469d2-scripts\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553491 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0e4fdb-6066-4163-9864-9e52443afdb0-combined-ca-bundle\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553509 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-lib\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553529 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7688c\" (UniqueName: \"kubernetes.io/projected/89539b2e-806f-4062-85fe-efac9c4469d2-kube-api-access-7688c\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553551 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa0e4fdb-6066-4163-9864-9e52443afdb0-scripts\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.553574 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa0e4fdb-6066-4163-9864-9e52443afdb0-ovn-controller-tls-certs\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.554547 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-run-ovn\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.554671 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-log-ovn\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.554806 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa0e4fdb-6066-4163-9864-9e52443afdb0-var-run\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.554814 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-log\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.554871 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-run\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.555009 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-etc-ovs\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.555123 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89539b2e-806f-4062-85fe-efac9c4469d2-var-lib\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.556985 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89539b2e-806f-4062-85fe-efac9c4469d2-scripts\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.557147 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa0e4fdb-6066-4163-9864-9e52443afdb0-scripts\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.559366 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0e4fdb-6066-4163-9864-9e52443afdb0-combined-ca-bundle\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.559394 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa0e4fdb-6066-4163-9864-9e52443afdb0-ovn-controller-tls-certs\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.582434 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8st6z\" (UniqueName: \"kubernetes.io/projected/aa0e4fdb-6066-4163-9864-9e52443afdb0-kube-api-access-8st6z\") pod \"ovn-controller-82r49\" (UID: \"aa0e4fdb-6066-4163-9864-9e52443afdb0\") " pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.582741 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7688c\" (UniqueName: \"kubernetes.io/projected/89539b2e-806f-4062-85fe-efac9c4469d2-kube-api-access-7688c\") pod \"ovn-controller-ovs-6l2z4\" (UID: \"89539b2e-806f-4062-85fe-efac9c4469d2\") " pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.616710 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-82r49" Nov 24 19:31:19 crc kubenswrapper[5035]: I1124 19:31:19.643903 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.380894 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.385388 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.390771 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.391549 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-ml58q" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.391964 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.395955 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.406270 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513551 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513634 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513692 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513714 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513769 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513793 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513837 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghvfp\" (UniqueName: \"kubernetes.io/projected/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-kube-api-access-ghvfp\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.513859 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.614923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghvfp\" (UniqueName: \"kubernetes.io/projected/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-kube-api-access-ghvfp\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.614992 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.615078 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.615129 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.615195 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.615212 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.615442 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.615462 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.615691 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.616414 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.616496 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.616706 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.621128 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.625891 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.628345 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.631528 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghvfp\" (UniqueName: \"kubernetes.io/projected/1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf-kube-api-access-ghvfp\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.643233 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf\") " pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:22 crc kubenswrapper[5035]: I1124 19:31:22.724143 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:24 crc kubenswrapper[5035]: I1124 19:31:24.334830 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 19:31:24 crc kubenswrapper[5035]: W1124 19:31:24.743180 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode665c207_0fb4_4876_89d1_cb18eadb80af.slice/crio-a31ee8bd90f8d3fdd7f04168a93603972df1a89b4c8394f3daa290d5014ac1d2 WatchSource:0}: Error finding container a31ee8bd90f8d3fdd7f04168a93603972df1a89b4c8394f3daa290d5014ac1d2: Status 404 returned error can't find the container with id a31ee8bd90f8d3fdd7f04168a93603972df1a89b4c8394f3daa290d5014ac1d2 Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.769948 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.770352 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pzdmt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bdd77c89-ndp97_openstack(5ab18ea7-92bb-4693-a40c-148f07c2862f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.772331 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" podUID="5ab18ea7-92bb-4693-a40c-148f07c2862f" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.777658 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.777798 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lmw2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7c6d9948dc-dqv6q_openstack(15c17b6f-1037-43de-8e7a-cf86311cc674): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.779064 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.818478 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.818670 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wqxdz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6584b49599-bwjn6_openstack(b56b4d33-ed61-434e-87bb-e0ee1e53513a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.818984 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.819067 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g9xnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6486446b9f-2cczt_openstack(1ea95d1a-843d-4a31-8c0d-8f351d47c0bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.820512 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" Nov 24 19:31:24 crc kubenswrapper[5035]: E1124 19:31:24.820570 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6584b49599-bwjn6" podUID="b56b4d33-ed61-434e-87bb-e0ee1e53513a" Nov 24 19:31:25 crc kubenswrapper[5035]: I1124 19:31:25.140917 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e665c207-0fb4-4876-89d1-cb18eadb80af","Type":"ContainerStarted","Data":"a31ee8bd90f8d3fdd7f04168a93603972df1a89b4c8394f3daa290d5014ac1d2"} Nov 24 19:31:25 crc kubenswrapper[5035]: E1124 19:31:25.142536 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" Nov 24 19:31:25 crc kubenswrapper[5035]: E1124 19:31:25.143588 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" Nov 24 19:31:25 crc kubenswrapper[5035]: I1124 19:31:25.251671 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:31:25 crc kubenswrapper[5035]: I1124 19:31:25.278822 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 19:31:25 crc kubenswrapper[5035]: W1124 19:31:25.287788 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62cb1b72_4c4f_4c1e_bf23_0dc8a318560d.slice/crio-b37449e7468f890bf9cbc3fa2a9e1dde62f41407731f092b8fd9a5b2649e3a9c WatchSource:0}: Error finding container b37449e7468f890bf9cbc3fa2a9e1dde62f41407731f092b8fd9a5b2649e3a9c: Status 404 returned error can't find the container with id b37449e7468f890bf9cbc3fa2a9e1dde62f41407731f092b8fd9a5b2649e3a9c Nov 24 19:31:25 crc kubenswrapper[5035]: I1124 19:31:25.593204 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-82r49"] Nov 24 19:31:25 crc kubenswrapper[5035]: I1124 19:31:25.598337 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 19:31:25 crc kubenswrapper[5035]: I1124 19:31:25.700089 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 19:31:25 crc kubenswrapper[5035]: W1124 19:31:25.714357 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ddbd9ab_f1f3_4243_be2d_6653c8dca6cf.slice/crio-ccdfc77c97eedb1012f7c9c7ae26ba0b9b7935e8c119e21ed78d19c46c5885a4 WatchSource:0}: Error finding container ccdfc77c97eedb1012f7c9c7ae26ba0b9b7935e8c119e21ed78d19c46c5885a4: Status 404 returned error can't find the container with id ccdfc77c97eedb1012f7c9c7ae26ba0b9b7935e8c119e21ed78d19c46c5885a4 Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.030720 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.036594 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.153640 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-bwjn6" event={"ID":"b56b4d33-ed61-434e-87bb-e0ee1e53513a","Type":"ContainerDied","Data":"5ad9670adc12b83d1e8b591aa349aef515885aad88ad323eb14d6d716fb42f93"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.153665 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-bwjn6" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.155201 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d","Type":"ContainerStarted","Data":"b37449e7468f890bf9cbc3fa2a9e1dde62f41407731f092b8fd9a5b2649e3a9c"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.157574 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cc0b947e-0136-4ccf-8203-21604c41ff7d","Type":"ContainerStarted","Data":"75477799fff1b06c66d227e174b2ddc0ed65206bd3f1b3ca09e53c6aed6fa773"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.159273 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ed0e7cb-164f-4689-8714-1f11bfa25725","Type":"ContainerStarted","Data":"13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.161863 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf","Type":"ContainerStarted","Data":"ccdfc77c97eedb1012f7c9c7ae26ba0b9b7935e8c119e21ed78d19c46c5885a4"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.163283 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fcbe1807-d08b-4d8a-9c92-954b8da6820d","Type":"ContainerStarted","Data":"a2194aefbfa2bd0c880a05c8ac393bd7240b7649c3c92043de46074b0d8e8069"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.164757 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-82r49" event={"ID":"aa0e4fdb-6066-4163-9864-9e52443afdb0","Type":"ContainerStarted","Data":"186edccea890075a13508faabbb3fedf4432ada1924e78034d3b4910ebab0849"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.166492 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8716e4c9-35b0-4922-afc2-13006d78283a","Type":"ContainerStarted","Data":"821d307dfb370f447c67d71297fcefc4e81c98065f7a613d6310b10ef3ad01a6"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.167471 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" event={"ID":"5ab18ea7-92bb-4693-a40c-148f07c2862f","Type":"ContainerDied","Data":"35132a7dbf2085eaee3b0235f75fcc4c5d58437432395b1b4d2536da5d82d577"} Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.167501 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-ndp97" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.176196 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqxdz\" (UniqueName: \"kubernetes.io/projected/b56b4d33-ed61-434e-87bb-e0ee1e53513a-kube-api-access-wqxdz\") pod \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.176265 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzdmt\" (UniqueName: \"kubernetes.io/projected/5ab18ea7-92bb-4693-a40c-148f07c2862f-kube-api-access-pzdmt\") pod \"5ab18ea7-92bb-4693-a40c-148f07c2862f\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.176375 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab18ea7-92bb-4693-a40c-148f07c2862f-config\") pod \"5ab18ea7-92bb-4693-a40c-148f07c2862f\" (UID: \"5ab18ea7-92bb-4693-a40c-148f07c2862f\") " Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.176489 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-dns-svc\") pod \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.176584 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-config\") pod \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\" (UID: \"b56b4d33-ed61-434e-87bb-e0ee1e53513a\") " Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.177667 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b56b4d33-ed61-434e-87bb-e0ee1e53513a" (UID: "b56b4d33-ed61-434e-87bb-e0ee1e53513a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.177715 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ab18ea7-92bb-4693-a40c-148f07c2862f-config" (OuterVolumeSpecName: "config") pod "5ab18ea7-92bb-4693-a40c-148f07c2862f" (UID: "5ab18ea7-92bb-4693-a40c-148f07c2862f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.178390 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-config" (OuterVolumeSpecName: "config") pod "b56b4d33-ed61-434e-87bb-e0ee1e53513a" (UID: "b56b4d33-ed61-434e-87bb-e0ee1e53513a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.182157 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab18ea7-92bb-4693-a40c-148f07c2862f-kube-api-access-pzdmt" (OuterVolumeSpecName: "kube-api-access-pzdmt") pod "5ab18ea7-92bb-4693-a40c-148f07c2862f" (UID: "5ab18ea7-92bb-4693-a40c-148f07c2862f"). InnerVolumeSpecName "kube-api-access-pzdmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.182763 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b56b4d33-ed61-434e-87bb-e0ee1e53513a-kube-api-access-wqxdz" (OuterVolumeSpecName: "kube-api-access-wqxdz") pod "b56b4d33-ed61-434e-87bb-e0ee1e53513a" (UID: "b56b4d33-ed61-434e-87bb-e0ee1e53513a"). InnerVolumeSpecName "kube-api-access-wqxdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.278576 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.278612 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b56b4d33-ed61-434e-87bb-e0ee1e53513a-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.278626 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqxdz\" (UniqueName: \"kubernetes.io/projected/b56b4d33-ed61-434e-87bb-e0ee1e53513a-kube-api-access-wqxdz\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.278638 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzdmt\" (UniqueName: \"kubernetes.io/projected/5ab18ea7-92bb-4693-a40c-148f07c2862f-kube-api-access-pzdmt\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.278648 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab18ea7-92bb-4693-a40c-148f07c2862f-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.356639 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6l2z4"] Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.494081 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-bwjn6"] Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.501976 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-bwjn6"] Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.522163 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-ndp97"] Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.534054 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-ndp97"] Nov 24 19:31:26 crc kubenswrapper[5035]: I1124 19:31:26.553606 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 19:31:26 crc kubenswrapper[5035]: W1124 19:31:26.941387 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74f6b29d_0844_4946_abaf_d331f8f07ba0.slice/crio-f5f96c6fba52047e03fe4615c5603294b89390ecacb4c8c6249136f2bc6bd7ca WatchSource:0}: Error finding container f5f96c6fba52047e03fe4615c5603294b89390ecacb4c8c6249136f2bc6bd7ca: Status 404 returned error can't find the container with id f5f96c6fba52047e03fe4615c5603294b89390ecacb4c8c6249136f2bc6bd7ca Nov 24 19:31:27 crc kubenswrapper[5035]: I1124 19:31:27.179013 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"74f6b29d-0844-4946-abaf-d331f8f07ba0","Type":"ContainerStarted","Data":"f5f96c6fba52047e03fe4615c5603294b89390ecacb4c8c6249136f2bc6bd7ca"} Nov 24 19:31:27 crc kubenswrapper[5035]: I1124 19:31:27.181738 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6l2z4" event={"ID":"89539b2e-806f-4062-85fe-efac9c4469d2","Type":"ContainerStarted","Data":"2b2b152820526d4913a08129987100cb9dfa604ba68023a22f1a9ddc339a37c2"} Nov 24 19:31:28 crc kubenswrapper[5035]: I1124 19:31:28.209753 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab18ea7-92bb-4693-a40c-148f07c2862f" path="/var/lib/kubelet/pods/5ab18ea7-92bb-4693-a40c-148f07c2862f/volumes" Nov 24 19:31:28 crc kubenswrapper[5035]: I1124 19:31:28.210747 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b56b4d33-ed61-434e-87bb-e0ee1e53513a" path="/var/lib/kubelet/pods/b56b4d33-ed61-434e-87bb-e0ee1e53513a/volumes" Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.214803 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf","Type":"ContainerStarted","Data":"66ea8b871aba99772fb3e22fff4f5988d1d2d25d8570693077caea5e24df662d"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.216332 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fcbe1807-d08b-4d8a-9c92-954b8da6820d","Type":"ContainerStarted","Data":"959495eb761729c3cb0d2d644f8553798dbfcec7cfd08304266444ed2ca05d58"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.217528 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d","Type":"ContainerStarted","Data":"35e7e0d133273a94df9a757beaa16dc7cc679231d933b0d86cddd240ce60be6a"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.219135 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"74f6b29d-0844-4946-abaf-d331f8f07ba0","Type":"ContainerStarted","Data":"4181c08a73d7cc58a1426918cbf63154a11b9683b052744612ad544417877cea"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.220084 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cc0b947e-0136-4ccf-8203-21604c41ff7d","Type":"ContainerStarted","Data":"f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.220685 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.222785 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-82r49" event={"ID":"aa0e4fdb-6066-4163-9864-9e52443afdb0","Type":"ContainerStarted","Data":"1a2706d18af9a2e14feecc8343b5b5e9c78f852d74eb387db8d01d511030ef54"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.222867 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-82r49" Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.224358 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6l2z4" event={"ID":"89539b2e-806f-4062-85fe-efac9c4469d2","Type":"ContainerStarted","Data":"34ee5c58bd287d976af1ddab8f924c1b36dd076e03f2fa2f91142e1539a46920"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.227214 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e665c207-0fb4-4876-89d1-cb18eadb80af","Type":"ContainerStarted","Data":"75b17f3def7b835fc3adc84817d4ffa785286cfbca45fb3a406e91c1e8c61b00"} Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.227590 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.400994 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=17.427202469 podStartE2EDuration="20.400970864s" podCreationTimestamp="2025-11-24 19:31:12 +0000 UTC" firstStartedPulling="2025-11-24 19:31:24.746278665 +0000 UTC m=+1023.268784922" lastFinishedPulling="2025-11-24 19:31:27.72004706 +0000 UTC m=+1026.242553317" observedRunningTime="2025-11-24 19:31:32.395619448 +0000 UTC m=+1030.918125705" watchObservedRunningTime="2025-11-24 19:31:32.400970864 +0000 UTC m=+1030.923477121" Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.417969 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-82r49" podStartSLOduration=7.408502083 podStartE2EDuration="13.417951309s" podCreationTimestamp="2025-11-24 19:31:19 +0000 UTC" firstStartedPulling="2025-11-24 19:31:25.662904346 +0000 UTC m=+1024.185410603" lastFinishedPulling="2025-11-24 19:31:31.672353572 +0000 UTC m=+1030.194859829" observedRunningTime="2025-11-24 19:31:32.417455376 +0000 UTC m=+1030.939961633" watchObservedRunningTime="2025-11-24 19:31:32.417951309 +0000 UTC m=+1030.940457566" Nov 24 19:31:32 crc kubenswrapper[5035]: I1124 19:31:32.461710 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.821286424 podStartE2EDuration="18.461684328s" podCreationTimestamp="2025-11-24 19:31:14 +0000 UTC" firstStartedPulling="2025-11-24 19:31:25.264011713 +0000 UTC m=+1023.786517970" lastFinishedPulling="2025-11-24 19:31:31.904409617 +0000 UTC m=+1030.426915874" observedRunningTime="2025-11-24 19:31:32.455725774 +0000 UTC m=+1030.978232031" watchObservedRunningTime="2025-11-24 19:31:32.461684328 +0000 UTC m=+1030.984190575" Nov 24 19:31:33 crc kubenswrapper[5035]: I1124 19:31:33.237553 5035 generic.go:334] "Generic (PLEG): container finished" podID="89539b2e-806f-4062-85fe-efac9c4469d2" containerID="34ee5c58bd287d976af1ddab8f924c1b36dd076e03f2fa2f91142e1539a46920" exitCode=0 Nov 24 19:31:33 crc kubenswrapper[5035]: I1124 19:31:33.238937 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6l2z4" event={"ID":"89539b2e-806f-4062-85fe-efac9c4469d2","Type":"ContainerDied","Data":"34ee5c58bd287d976af1ddab8f924c1b36dd076e03f2fa2f91142e1539a46920"} Nov 24 19:31:34 crc kubenswrapper[5035]: I1124 19:31:34.248210 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6l2z4" event={"ID":"89539b2e-806f-4062-85fe-efac9c4469d2","Type":"ContainerStarted","Data":"25d68337fb296ec6860dfb455366b0749e2833c44366389ba199ff5189d2304c"} Nov 24 19:31:34 crc kubenswrapper[5035]: I1124 19:31:34.248885 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:34 crc kubenswrapper[5035]: I1124 19:31:34.248902 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:31:34 crc kubenswrapper[5035]: I1124 19:31:34.248913 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6l2z4" event={"ID":"89539b2e-806f-4062-85fe-efac9c4469d2","Type":"ContainerStarted","Data":"99a0358d621f3759a324253864441de541ea0fd922263ad0f859414612956338"} Nov 24 19:31:34 crc kubenswrapper[5035]: I1124 19:31:34.277237 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6l2z4" podStartSLOduration=10.893037845 podStartE2EDuration="15.277212224s" podCreationTimestamp="2025-11-24 19:31:19 +0000 UTC" firstStartedPulling="2025-11-24 19:31:26.942534638 +0000 UTC m=+1025.465040895" lastFinishedPulling="2025-11-24 19:31:31.326709017 +0000 UTC m=+1029.849215274" observedRunningTime="2025-11-24 19:31:34.268532267 +0000 UTC m=+1032.791038534" watchObservedRunningTime="2025-11-24 19:31:34.277212224 +0000 UTC m=+1032.799718521" Nov 24 19:31:36 crc kubenswrapper[5035]: I1124 19:31:36.269913 5035 generic.go:334] "Generic (PLEG): container finished" podID="fcbe1807-d08b-4d8a-9c92-954b8da6820d" containerID="959495eb761729c3cb0d2d644f8553798dbfcec7cfd08304266444ed2ca05d58" exitCode=0 Nov 24 19:31:36 crc kubenswrapper[5035]: I1124 19:31:36.270452 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fcbe1807-d08b-4d8a-9c92-954b8da6820d","Type":"ContainerDied","Data":"959495eb761729c3cb0d2d644f8553798dbfcec7cfd08304266444ed2ca05d58"} Nov 24 19:31:36 crc kubenswrapper[5035]: I1124 19:31:36.279128 5035 generic.go:334] "Generic (PLEG): container finished" podID="62cb1b72-4c4f-4c1e-bf23-0dc8a318560d" containerID="35e7e0d133273a94df9a757beaa16dc7cc679231d933b0d86cddd240ce60be6a" exitCode=0 Nov 24 19:31:36 crc kubenswrapper[5035]: I1124 19:31:36.279170 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d","Type":"ContainerDied","Data":"35e7e0d133273a94df9a757beaa16dc7cc679231d933b0d86cddd240ce60be6a"} Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.291127 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf","Type":"ContainerStarted","Data":"854e3c0faf42c3283547d82aea6a22f0a9d39d4ac6898fdadb8235d2c8a47132"} Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.294143 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fcbe1807-d08b-4d8a-9c92-954b8da6820d","Type":"ContainerStarted","Data":"d5b6840146983c077bbd55275366ae5b8e8f041f76c18efb1e88b16ca38c52b2"} Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.296845 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"62cb1b72-4c4f-4c1e-bf23-0dc8a318560d","Type":"ContainerStarted","Data":"996aa0801d65fa63dea79e53be4ced14ffd5db03069af14b2de1c66bbd5d91f6"} Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.298906 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"74f6b29d-0844-4946-abaf-d331f8f07ba0","Type":"ContainerStarted","Data":"87eb94cc8beda6e6785e1a04a6e0e9263db02449298d6ed63be5ab9f91903a9b"} Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.337410 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.034316095 podStartE2EDuration="16.337389636s" podCreationTimestamp="2025-11-24 19:31:21 +0000 UTC" firstStartedPulling="2025-11-24 19:31:25.716942905 +0000 UTC m=+1024.239449162" lastFinishedPulling="2025-11-24 19:31:35.020016446 +0000 UTC m=+1033.542522703" observedRunningTime="2025-11-24 19:31:37.315531387 +0000 UTC m=+1035.838037654" watchObservedRunningTime="2025-11-24 19:31:37.337389636 +0000 UTC m=+1035.859895923" Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.338752 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=12.299690787 podStartE2EDuration="20.338744403s" podCreationTimestamp="2025-11-24 19:31:17 +0000 UTC" firstStartedPulling="2025-11-24 19:31:26.965170248 +0000 UTC m=+1025.487676515" lastFinishedPulling="2025-11-24 19:31:35.004223854 +0000 UTC m=+1033.526730131" observedRunningTime="2025-11-24 19:31:37.333061998 +0000 UTC m=+1035.855568265" watchObservedRunningTime="2025-11-24 19:31:37.338744403 +0000 UTC m=+1035.861250700" Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.358287 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.859178438 podStartE2EDuration="28.358266018s" podCreationTimestamp="2025-11-24 19:31:09 +0000 UTC" firstStartedPulling="2025-11-24 19:31:25.661690313 +0000 UTC m=+1024.184196570" lastFinishedPulling="2025-11-24 19:31:31.160777893 +0000 UTC m=+1029.683284150" observedRunningTime="2025-11-24 19:31:37.350033292 +0000 UTC m=+1035.872539579" watchObservedRunningTime="2025-11-24 19:31:37.358266018 +0000 UTC m=+1035.880772295" Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.377566 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.516432561 podStartE2EDuration="26.377549236s" podCreationTimestamp="2025-11-24 19:31:11 +0000 UTC" firstStartedPulling="2025-11-24 19:31:25.299690149 +0000 UTC m=+1023.822196406" lastFinishedPulling="2025-11-24 19:31:31.160806784 +0000 UTC m=+1029.683313081" observedRunningTime="2025-11-24 19:31:37.37405071 +0000 UTC m=+1035.896556967" watchObservedRunningTime="2025-11-24 19:31:37.377549236 +0000 UTC m=+1035.900055483" Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.724771 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.724809 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.790937 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:37 crc kubenswrapper[5035]: I1124 19:31:37.951597 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.307555 5035 generic.go:334] "Generic (PLEG): container finished" podID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerID="d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56" exitCode=0 Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.307660 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" event={"ID":"15c17b6f-1037-43de-8e7a-cf86311cc674","Type":"ContainerDied","Data":"d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56"} Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.310428 5035 generic.go:334] "Generic (PLEG): container finished" podID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerID="f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5" exitCode=0 Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.310522 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" event={"ID":"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc","Type":"ContainerDied","Data":"f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5"} Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.361571 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.624810 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-dqv6q"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.659198 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-lp2gx"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.660443 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.662513 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.670066 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.679000 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-rxj4x"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.680609 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.683227 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.690514 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-lp2gx"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.711029 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rxj4x"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803102 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803169 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-dns-svc\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803299 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-ovs-rundir\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803417 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-config\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803459 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-combined-ca-bundle\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803486 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-ovn-rundir\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803591 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-config\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803697 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-ovsdbserver-sb\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803723 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwwch\" (UniqueName: \"kubernetes.io/projected/4d254b9e-aa95-4296-95aa-52eacc395748-kube-api-access-wwwch\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.803775 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rfbd\" (UniqueName: \"kubernetes.io/projected/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-kube-api-access-9rfbd\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.905129 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-config\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.905197 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-combined-ca-bundle\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.905230 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-ovn-rundir\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.905259 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-config\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.905350 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-ovsdbserver-sb\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.905666 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-ovn-rundir\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.906387 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-config\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.905489 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwwch\" (UniqueName: \"kubernetes.io/projected/4d254b9e-aa95-4296-95aa-52eacc395748-kube-api-access-wwwch\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.906490 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rfbd\" (UniqueName: \"kubernetes.io/projected/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-kube-api-access-9rfbd\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.906487 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-config\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.906582 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.906770 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-dns-svc\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.906832 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-ovs-rundir\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.907018 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-ovs-rundir\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.907046 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-ovsdbserver-sb\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.907823 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-dns-svc\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.911226 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-combined-ca-bundle\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.911600 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.918285 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2cczt"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.929788 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rfbd\" (UniqueName: \"kubernetes.io/projected/2a4176a5-5ca8-4a71-8cab-3330b6af69f5-kube-api-access-9rfbd\") pod \"ovn-controller-metrics-rxj4x\" (UID: \"2a4176a5-5ca8-4a71-8cab-3330b6af69f5\") " pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.931542 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwwch\" (UniqueName: \"kubernetes.io/projected/4d254b9e-aa95-4296-95aa-52eacc395748-kube-api-access-wwwch\") pod \"dnsmasq-dns-5848494dd9-lp2gx\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.946814 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-7vcb7"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.948444 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.950339 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.960116 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-7vcb7"] Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.982804 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:38 crc kubenswrapper[5035]: I1124 19:31:38.995804 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rxj4x" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.109313 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.109365 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-config\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.109426 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-dns-svc\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.109617 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwgxf\" (UniqueName: \"kubernetes.io/projected/4875d758-6b35-47c1-8b60-80a8d718fdbd-kube-api-access-gwgxf\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.109820 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.211665 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.212008 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.212052 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-config\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.212121 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-dns-svc\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.212171 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwgxf\" (UniqueName: \"kubernetes.io/projected/4875d758-6b35-47c1-8b60-80a8d718fdbd-kube-api-access-gwgxf\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.213266 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.215960 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-config\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.216049 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.216264 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-dns-svc\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.229441 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwgxf\" (UniqueName: \"kubernetes.io/projected/4875d758-6b35-47c1-8b60-80a8d718fdbd-kube-api-access-gwgxf\") pod \"dnsmasq-dns-5c7b6b5695-7vcb7\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.321209 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" event={"ID":"15c17b6f-1037-43de-8e7a-cf86311cc674","Type":"ContainerStarted","Data":"8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534"} Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.321358 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerName="dnsmasq-dns" containerID="cri-o://8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534" gracePeriod=10 Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.321457 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.324146 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerName="dnsmasq-dns" containerID="cri-o://baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868" gracePeriod=10 Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.324238 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" event={"ID":"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc","Type":"ContainerStarted","Data":"baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868"} Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.324810 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.340399 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" podStartSLOduration=2.45045643 podStartE2EDuration="31.340382236s" podCreationTimestamp="2025-11-24 19:31:08 +0000 UTC" firstStartedPulling="2025-11-24 19:31:08.934175459 +0000 UTC m=+1007.456681716" lastFinishedPulling="2025-11-24 19:31:37.824101225 +0000 UTC m=+1036.346607522" observedRunningTime="2025-11-24 19:31:39.338097404 +0000 UTC m=+1037.860603661" watchObservedRunningTime="2025-11-24 19:31:39.340382236 +0000 UTC m=+1037.862888493" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.360231 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" podStartSLOduration=2.835865446 podStartE2EDuration="31.36021589s" podCreationTimestamp="2025-11-24 19:31:08 +0000 UTC" firstStartedPulling="2025-11-24 19:31:09.083835547 +0000 UTC m=+1007.606341804" lastFinishedPulling="2025-11-24 19:31:37.608185991 +0000 UTC m=+1036.130692248" observedRunningTime="2025-11-24 19:31:39.35440569 +0000 UTC m=+1037.876911947" watchObservedRunningTime="2025-11-24 19:31:39.36021589 +0000 UTC m=+1037.882722137" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.379619 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.448322 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rxj4x"] Nov 24 19:31:39 crc kubenswrapper[5035]: W1124 19:31:39.464588 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a4176a5_5ca8_4a71_8cab_3330b6af69f5.slice/crio-e17b8aab58af817a6af9a824eb43c7d693bcecba4ce5b777bf47e73ac4ac0e3b WatchSource:0}: Error finding container e17b8aab58af817a6af9a824eb43c7d693bcecba4ce5b777bf47e73ac4ac0e3b: Status 404 returned error can't find the container with id e17b8aab58af817a6af9a824eb43c7d693bcecba4ce5b777bf47e73ac4ac0e3b Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.524653 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-lp2gx"] Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.670527 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.717595 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:39 crc kubenswrapper[5035]: I1124 19:31:39.872780 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-7vcb7"] Nov 24 19:31:39 crc kubenswrapper[5035]: W1124 19:31:39.900439 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4875d758_6b35_47c1_8b60_80a8d718fdbd.slice/crio-1e154e6a4e2d4a3aa5d2bc56baded04402ee48c3d4cc46a068cbbe7038235e3f WatchSource:0}: Error finding container 1e154e6a4e2d4a3aa5d2bc56baded04402ee48c3d4cc46a068cbbe7038235e3f: Status 404 returned error can't find the container with id 1e154e6a4e2d4a3aa5d2bc56baded04402ee48c3d4cc46a068cbbe7038235e3f Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.086709 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.094986 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.240864 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-dns-svc\") pod \"15c17b6f-1037-43de-8e7a-cf86311cc674\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.240929 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-config\") pod \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.240963 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9xnw\" (UniqueName: \"kubernetes.io/projected/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-kube-api-access-g9xnw\") pod \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.240990 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-config\") pod \"15c17b6f-1037-43de-8e7a-cf86311cc674\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.241030 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmw2j\" (UniqueName: \"kubernetes.io/projected/15c17b6f-1037-43de-8e7a-cf86311cc674-kube-api-access-lmw2j\") pod \"15c17b6f-1037-43de-8e7a-cf86311cc674\" (UID: \"15c17b6f-1037-43de-8e7a-cf86311cc674\") " Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.241114 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-dns-svc\") pod \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\" (UID: \"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc\") " Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.246510 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-kube-api-access-g9xnw" (OuterVolumeSpecName: "kube-api-access-g9xnw") pod "1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" (UID: "1ea95d1a-843d-4a31-8c0d-8f351d47c0bc"). InnerVolumeSpecName "kube-api-access-g9xnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.249183 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c17b6f-1037-43de-8e7a-cf86311cc674-kube-api-access-lmw2j" (OuterVolumeSpecName: "kube-api-access-lmw2j") pod "15c17b6f-1037-43de-8e7a-cf86311cc674" (UID: "15c17b6f-1037-43de-8e7a-cf86311cc674"). InnerVolumeSpecName "kube-api-access-lmw2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.277519 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-config" (OuterVolumeSpecName: "config") pod "15c17b6f-1037-43de-8e7a-cf86311cc674" (UID: "15c17b6f-1037-43de-8e7a-cf86311cc674"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.285818 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15c17b6f-1037-43de-8e7a-cf86311cc674" (UID: "15c17b6f-1037-43de-8e7a-cf86311cc674"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.288838 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-config" (OuterVolumeSpecName: "config") pod "1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" (UID: "1ea95d1a-843d-4a31-8c0d-8f351d47c0bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.291550 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" (UID: "1ea95d1a-843d-4a31-8c0d-8f351d47c0bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.332385 5035 generic.go:334] "Generic (PLEG): container finished" podID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerID="baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868" exitCode=0 Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.332453 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.332466 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" event={"ID":"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc","Type":"ContainerDied","Data":"baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.332492 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-2cczt" event={"ID":"1ea95d1a-843d-4a31-8c0d-8f351d47c0bc","Type":"ContainerDied","Data":"259e30c6b7794ed4af98c767a6d24e3555135383f2f9e46eede2d9a35c2d9bd5"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.332507 5035 scope.go:117] "RemoveContainer" containerID="baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.334253 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rxj4x" event={"ID":"2a4176a5-5ca8-4a71-8cab-3330b6af69f5","Type":"ContainerStarted","Data":"b87ed502651290123146adbbb71406ea71f5cec8fd444cf01689ce584361aea3"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.334282 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rxj4x" event={"ID":"2a4176a5-5ca8-4a71-8cab-3330b6af69f5","Type":"ContainerStarted","Data":"e17b8aab58af817a6af9a824eb43c7d693bcecba4ce5b777bf47e73ac4ac0e3b"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.335720 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d254b9e-aa95-4296-95aa-52eacc395748" containerID="8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235" exitCode=0 Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.336305 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" event={"ID":"4d254b9e-aa95-4296-95aa-52eacc395748","Type":"ContainerDied","Data":"8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.336347 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" event={"ID":"4d254b9e-aa95-4296-95aa-52eacc395748","Type":"ContainerStarted","Data":"06ded05bee98d643c4a38c0ebcfd99288a2054663f1725a28140567c9702f46f"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.344161 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.345260 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.349350 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.349651 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9xnw\" (UniqueName: \"kubernetes.io/projected/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc-kube-api-access-g9xnw\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.349745 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c17b6f-1037-43de-8e7a-cf86311cc674-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.349835 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmw2j\" (UniqueName: \"kubernetes.io/projected/15c17b6f-1037-43de-8e7a-cf86311cc674-kube-api-access-lmw2j\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.349830 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" event={"ID":"15c17b6f-1037-43de-8e7a-cf86311cc674","Type":"ContainerDied","Data":"8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.349799 5035 generic.go:334] "Generic (PLEG): container finished" podID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerID="8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534" exitCode=0 Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.350128 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" event={"ID":"15c17b6f-1037-43de-8e7a-cf86311cc674","Type":"ContainerDied","Data":"acaef67ec5e5ead8913323a556b0dc04cece24b3bca342b7059812df061a6369"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.350354 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-dqv6q" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.355094 5035 generic.go:334] "Generic (PLEG): container finished" podID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerID="0a3e0fc8da37a85db5fb228b02cc23c13ba574252793f3febca5965fcd702f2a" exitCode=0 Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.355458 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" event={"ID":"4875d758-6b35-47c1-8b60-80a8d718fdbd","Type":"ContainerDied","Data":"0a3e0fc8da37a85db5fb228b02cc23c13ba574252793f3febca5965fcd702f2a"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.355538 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" event={"ID":"4875d758-6b35-47c1-8b60-80a8d718fdbd","Type":"ContainerStarted","Data":"1e154e6a4e2d4a3aa5d2bc56baded04402ee48c3d4cc46a068cbbe7038235e3f"} Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.361214 5035 scope.go:117] "RemoveContainer" containerID="f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.404896 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-rxj4x" podStartSLOduration=2.404876018 podStartE2EDuration="2.404876018s" podCreationTimestamp="2025-11-24 19:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:31:40.356915134 +0000 UTC m=+1038.879421391" watchObservedRunningTime="2025-11-24 19:31:40.404876018 +0000 UTC m=+1038.927382265" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.426986 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.427066 5035 scope.go:117] "RemoveContainer" containerID="baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868" Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.427435 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868\": container with ID starting with baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868 not found: ID does not exist" containerID="baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.427485 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868"} err="failed to get container status \"baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868\": rpc error: code = NotFound desc = could not find container \"baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868\": container with ID starting with baef9b0a59c90e624b627ac221b99269952bfcdde4f17fc3f2df9b5a08a21868 not found: ID does not exist" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.427527 5035 scope.go:117] "RemoveContainer" containerID="f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5" Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.427842 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5\": container with ID starting with f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5 not found: ID does not exist" containerID="f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.427876 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5"} err="failed to get container status \"f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5\": rpc error: code = NotFound desc = could not find container \"f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5\": container with ID starting with f164e4d4643359656dd74d793b427678ca4535fc901c05a1865dd4ef5f6b24d5 not found: ID does not exist" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.427895 5035 scope.go:117] "RemoveContainer" containerID="8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.445059 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2cczt"] Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.450483 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-2cczt"] Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.472458 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-dqv6q"] Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.484908 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-dqv6q"] Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.495160 5035 scope.go:117] "RemoveContainer" containerID="d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.536754 5035 scope.go:117] "RemoveContainer" containerID="8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534" Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.537512 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534\": container with ID starting with 8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534 not found: ID does not exist" containerID="8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.537573 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534"} err="failed to get container status \"8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534\": rpc error: code = NotFound desc = could not find container \"8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534\": container with ID starting with 8a99078b591488a13e3529e9e554c1fa735d2cff3b0c10c656026f64e305b534 not found: ID does not exist" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.537605 5035 scope.go:117] "RemoveContainer" containerID="d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56" Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.538433 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56\": container with ID starting with d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56 not found: ID does not exist" containerID="d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.538473 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56"} err="failed to get container status \"d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56\": rpc error: code = NotFound desc = could not find container \"d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56\": container with ID starting with d7af24f9554c0960b86290b032d51551076efa8c40e3f80fa827cf6c2b751e56 not found: ID does not exist" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.621214 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.621562 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerName="dnsmasq-dns" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.621580 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerName="dnsmasq-dns" Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.621591 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerName="init" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.621598 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerName="init" Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.621610 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerName="init" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.621615 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerName="init" Nov 24 19:31:40 crc kubenswrapper[5035]: E1124 19:31:40.621638 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerName="dnsmasq-dns" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.621644 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerName="dnsmasq-dns" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.621789 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" containerName="dnsmasq-dns" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.621804 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" containerName="dnsmasq-dns" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.629996 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.633885 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.633967 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.634028 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8hxcr" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.635974 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.636374 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.757803 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.757924 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9637335d-6e3c-4f1a-bfcd-a081f20e8226-config\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.757957 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7z74\" (UniqueName: \"kubernetes.io/projected/9637335d-6e3c-4f1a-bfcd-a081f20e8226-kube-api-access-m7z74\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.758012 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.758080 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9637335d-6e3c-4f1a-bfcd-a081f20e8226-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.758141 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9637335d-6e3c-4f1a-bfcd-a081f20e8226-scripts\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.758217 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.860251 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.860326 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9637335d-6e3c-4f1a-bfcd-a081f20e8226-config\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.860353 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7z74\" (UniqueName: \"kubernetes.io/projected/9637335d-6e3c-4f1a-bfcd-a081f20e8226-kube-api-access-m7z74\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.860395 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.860415 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9637335d-6e3c-4f1a-bfcd-a081f20e8226-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.860438 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9637335d-6e3c-4f1a-bfcd-a081f20e8226-scripts\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.860471 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.861577 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9637335d-6e3c-4f1a-bfcd-a081f20e8226-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.861897 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9637335d-6e3c-4f1a-bfcd-a081f20e8226-config\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.862086 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9637335d-6e3c-4f1a-bfcd-a081f20e8226-scripts\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.865976 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.866155 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.885148 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9637335d-6e3c-4f1a-bfcd-a081f20e8226-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:40 crc kubenswrapper[5035]: I1124 19:31:40.890233 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7z74\" (UniqueName: \"kubernetes.io/projected/9637335d-6e3c-4f1a-bfcd-a081f20e8226-kube-api-access-m7z74\") pod \"ovn-northd-0\" (UID: \"9637335d-6e3c-4f1a-bfcd-a081f20e8226\") " pod="openstack/ovn-northd-0" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.016801 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.351316 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.351364 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.367127 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" event={"ID":"4875d758-6b35-47c1-8b60-80a8d718fdbd","Type":"ContainerStarted","Data":"0e9ed067c749b72e78ae141bde292b7b156ab3d786b8866e1a565dfd19a5ce7c"} Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.368322 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.371639 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" event={"ID":"4d254b9e-aa95-4296-95aa-52eacc395748","Type":"ContainerStarted","Data":"794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272"} Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.371767 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.382838 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" podStartSLOduration=3.382822038 podStartE2EDuration="3.382822038s" podCreationTimestamp="2025-11-24 19:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:31:41.381381488 +0000 UTC m=+1039.903887745" watchObservedRunningTime="2025-11-24 19:31:41.382822038 +0000 UTC m=+1039.905328295" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.405955 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" podStartSLOduration=3.405934641 podStartE2EDuration="3.405934641s" podCreationTimestamp="2025-11-24 19:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:31:41.399733711 +0000 UTC m=+1039.922239978" watchObservedRunningTime="2025-11-24 19:31:41.405934641 +0000 UTC m=+1039.928440908" Nov 24 19:31:41 crc kubenswrapper[5035]: I1124 19:31:41.464614 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 19:31:41 crc kubenswrapper[5035]: W1124 19:31:41.469490 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9637335d_6e3c_4f1a_bfcd_a081f20e8226.slice/crio-90b28c8b89842b3e0a384274e1fbb70e0d8029bd083f14be2d7b33e2cd00ab3d WatchSource:0}: Error finding container 90b28c8b89842b3e0a384274e1fbb70e0d8029bd083f14be2d7b33e2cd00ab3d: Status 404 returned error can't find the container with id 90b28c8b89842b3e0a384274e1fbb70e0d8029bd083f14be2d7b33e2cd00ab3d Nov 24 19:31:42 crc kubenswrapper[5035]: I1124 19:31:42.216235 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15c17b6f-1037-43de-8e7a-cf86311cc674" path="/var/lib/kubelet/pods/15c17b6f-1037-43de-8e7a-cf86311cc674/volumes" Nov 24 19:31:42 crc kubenswrapper[5035]: I1124 19:31:42.217804 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea95d1a-843d-4a31-8c0d-8f351d47c0bc" path="/var/lib/kubelet/pods/1ea95d1a-843d-4a31-8c0d-8f351d47c0bc/volumes" Nov 24 19:31:42 crc kubenswrapper[5035]: I1124 19:31:42.385905 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9637335d-6e3c-4f1a-bfcd-a081f20e8226","Type":"ContainerStarted","Data":"90b28c8b89842b3e0a384274e1fbb70e0d8029bd083f14be2d7b33e2cd00ab3d"} Nov 24 19:31:42 crc kubenswrapper[5035]: I1124 19:31:42.668478 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:42 crc kubenswrapper[5035]: I1124 19:31:42.668764 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:42 crc kubenswrapper[5035]: I1124 19:31:42.874121 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 19:31:43 crc kubenswrapper[5035]: I1124 19:31:43.139550 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 19:31:44 crc kubenswrapper[5035]: I1124 19:31:44.698171 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.223761 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-edc5-account-create-bfvds"] Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.225248 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.227218 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.239671 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-edc5-account-create-bfvds"] Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.247847 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pmbwf"] Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.249175 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.265441 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pmbwf"] Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.287251 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a40bddb-4969-4100-a311-1fd2614bd9f8-operator-scripts\") pod \"glance-edc5-account-create-bfvds\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.287362 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395eef76-f4f1-4543-84b0-0d69abb77653-operator-scripts\") pod \"glance-db-create-pmbwf\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.287510 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6sjb\" (UniqueName: \"kubernetes.io/projected/1a40bddb-4969-4100-a311-1fd2614bd9f8-kube-api-access-z6sjb\") pod \"glance-edc5-account-create-bfvds\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.287532 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb7p4\" (UniqueName: \"kubernetes.io/projected/395eef76-f4f1-4543-84b0-0d69abb77653-kube-api-access-rb7p4\") pod \"glance-db-create-pmbwf\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.388445 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6sjb\" (UniqueName: \"kubernetes.io/projected/1a40bddb-4969-4100-a311-1fd2614bd9f8-kube-api-access-z6sjb\") pod \"glance-edc5-account-create-bfvds\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.388799 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb7p4\" (UniqueName: \"kubernetes.io/projected/395eef76-f4f1-4543-84b0-0d69abb77653-kube-api-access-rb7p4\") pod \"glance-db-create-pmbwf\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.388899 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a40bddb-4969-4100-a311-1fd2614bd9f8-operator-scripts\") pod \"glance-edc5-account-create-bfvds\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.388958 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395eef76-f4f1-4543-84b0-0d69abb77653-operator-scripts\") pod \"glance-db-create-pmbwf\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.390062 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a40bddb-4969-4100-a311-1fd2614bd9f8-operator-scripts\") pod \"glance-edc5-account-create-bfvds\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.390214 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395eef76-f4f1-4543-84b0-0d69abb77653-operator-scripts\") pod \"glance-db-create-pmbwf\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.407508 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb7p4\" (UniqueName: \"kubernetes.io/projected/395eef76-f4f1-4543-84b0-0d69abb77653-kube-api-access-rb7p4\") pod \"glance-db-create-pmbwf\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.415017 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6sjb\" (UniqueName: \"kubernetes.io/projected/1a40bddb-4969-4100-a311-1fd2614bd9f8-kube-api-access-z6sjb\") pod \"glance-edc5-account-create-bfvds\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.545641 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.566349 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:48 crc kubenswrapper[5035]: I1124 19:31:48.985491 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.034121 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.046273 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-edc5-account-create-bfvds"] Nov 24 19:31:49 crc kubenswrapper[5035]: W1124 19:31:49.088107 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a40bddb_4969_4100_a311_1fd2614bd9f8.slice/crio-5a9a06773df51d39f3d63e8be454521b00e4f0d7e3087054c6982f56d64d3356 WatchSource:0}: Error finding container 5a9a06773df51d39f3d63e8be454521b00e4f0d7e3087054c6982f56d64d3356: Status 404 returned error can't find the container with id 5a9a06773df51d39f3d63e8be454521b00e4f0d7e3087054c6982f56d64d3356 Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.133396 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pmbwf"] Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.162159 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.381565 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.444814 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-lp2gx"] Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.447970 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9637335d-6e3c-4f1a-bfcd-a081f20e8226","Type":"ContainerStarted","Data":"dc7acb14e611f807d291572e813ebfe358f7f7ed0a7f9e75ac4055e1683e8ef5"} Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.452988 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-edc5-account-create-bfvds" event={"ID":"1a40bddb-4969-4100-a311-1fd2614bd9f8","Type":"ContainerStarted","Data":"0137522e2f0147317312b1acc52d8f5f9815e67b3584dbb91490d969b14317fb"} Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.453030 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-edc5-account-create-bfvds" event={"ID":"1a40bddb-4969-4100-a311-1fd2614bd9f8","Type":"ContainerStarted","Data":"5a9a06773df51d39f3d63e8be454521b00e4f0d7e3087054c6982f56d64d3356"} Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.456021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmbwf" event={"ID":"395eef76-f4f1-4543-84b0-0d69abb77653","Type":"ContainerStarted","Data":"a139d5c50022bea46fcaabbf66a045cfd73e2839694412004c8747dfe942cf21"} Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.456067 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmbwf" event={"ID":"395eef76-f4f1-4543-84b0-0d69abb77653","Type":"ContainerStarted","Data":"9de59e1249e0ec0c6fdb8cae76ce58cedc0388f766bc0d23d0993808c89355ca"} Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.456132 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" podUID="4d254b9e-aa95-4296-95aa-52eacc395748" containerName="dnsmasq-dns" containerID="cri-o://794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272" gracePeriod=10 Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.478923 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-edc5-account-create-bfvds" podStartSLOduration=1.478900555 podStartE2EDuration="1.478900555s" podCreationTimestamp="2025-11-24 19:31:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:31:49.471221805 +0000 UTC m=+1047.993728062" watchObservedRunningTime="2025-11-24 19:31:49.478900555 +0000 UTC m=+1048.001406802" Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.506482 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-pmbwf" podStartSLOduration=1.506459599 podStartE2EDuration="1.506459599s" podCreationTimestamp="2025-11-24 19:31:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:31:49.499693454 +0000 UTC m=+1048.022199711" watchObservedRunningTime="2025-11-24 19:31:49.506459599 +0000 UTC m=+1048.028965856" Nov 24 19:31:49 crc kubenswrapper[5035]: I1124 19:31:49.907936 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.018558 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-config\") pod \"4d254b9e-aa95-4296-95aa-52eacc395748\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.018650 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-ovsdbserver-sb\") pod \"4d254b9e-aa95-4296-95aa-52eacc395748\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.018676 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-dns-svc\") pod \"4d254b9e-aa95-4296-95aa-52eacc395748\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.018860 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwwch\" (UniqueName: \"kubernetes.io/projected/4d254b9e-aa95-4296-95aa-52eacc395748-kube-api-access-wwwch\") pod \"4d254b9e-aa95-4296-95aa-52eacc395748\" (UID: \"4d254b9e-aa95-4296-95aa-52eacc395748\") " Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.039517 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d254b9e-aa95-4296-95aa-52eacc395748-kube-api-access-wwwch" (OuterVolumeSpecName: "kube-api-access-wwwch") pod "4d254b9e-aa95-4296-95aa-52eacc395748" (UID: "4d254b9e-aa95-4296-95aa-52eacc395748"). InnerVolumeSpecName "kube-api-access-wwwch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.092135 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d254b9e-aa95-4296-95aa-52eacc395748" (UID: "4d254b9e-aa95-4296-95aa-52eacc395748"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.100994 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-config" (OuterVolumeSpecName: "config") pod "4d254b9e-aa95-4296-95aa-52eacc395748" (UID: "4d254b9e-aa95-4296-95aa-52eacc395748"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.102799 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4d254b9e-aa95-4296-95aa-52eacc395748" (UID: "4d254b9e-aa95-4296-95aa-52eacc395748"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.121377 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwwch\" (UniqueName: \"kubernetes.io/projected/4d254b9e-aa95-4296-95aa-52eacc395748-kube-api-access-wwwch\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.121404 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.121414 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.121422 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d254b9e-aa95-4296-95aa-52eacc395748-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.465170 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9637335d-6e3c-4f1a-bfcd-a081f20e8226","Type":"ContainerStarted","Data":"961561d68e8d7a795cd61adf0247b852e640a3934eef556b17a1bebdd47be11d"} Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.465280 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.466694 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d254b9e-aa95-4296-95aa-52eacc395748" containerID="794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272" exitCode=0 Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.466782 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" event={"ID":"4d254b9e-aa95-4296-95aa-52eacc395748","Type":"ContainerDied","Data":"794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272"} Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.466806 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" event={"ID":"4d254b9e-aa95-4296-95aa-52eacc395748","Type":"ContainerDied","Data":"06ded05bee98d643c4a38c0ebcfd99288a2054663f1725a28140567c9702f46f"} Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.466822 5035 scope.go:117] "RemoveContainer" containerID="794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.466849 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-lp2gx" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.475959 5035 generic.go:334] "Generic (PLEG): container finished" podID="1a40bddb-4969-4100-a311-1fd2614bd9f8" containerID="0137522e2f0147317312b1acc52d8f5f9815e67b3584dbb91490d969b14317fb" exitCode=0 Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.476000 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-edc5-account-create-bfvds" event={"ID":"1a40bddb-4969-4100-a311-1fd2614bd9f8","Type":"ContainerDied","Data":"0137522e2f0147317312b1acc52d8f5f9815e67b3584dbb91490d969b14317fb"} Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.486361 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.8625686200000002 podStartE2EDuration="10.486332293s" podCreationTimestamp="2025-11-24 19:31:40 +0000 UTC" firstStartedPulling="2025-11-24 19:31:41.473907222 +0000 UTC m=+1039.996413479" lastFinishedPulling="2025-11-24 19:31:49.097670885 +0000 UTC m=+1047.620177152" observedRunningTime="2025-11-24 19:31:50.484563104 +0000 UTC m=+1049.007069371" watchObservedRunningTime="2025-11-24 19:31:50.486332293 +0000 UTC m=+1049.008838580" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.486735 5035 generic.go:334] "Generic (PLEG): container finished" podID="395eef76-f4f1-4543-84b0-0d69abb77653" containerID="a139d5c50022bea46fcaabbf66a045cfd73e2839694412004c8747dfe942cf21" exitCode=0 Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.486782 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmbwf" event={"ID":"395eef76-f4f1-4543-84b0-0d69abb77653","Type":"ContainerDied","Data":"a139d5c50022bea46fcaabbf66a045cfd73e2839694412004c8747dfe942cf21"} Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.504077 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-lp2gx"] Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.510958 5035 scope.go:117] "RemoveContainer" containerID="8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.512754 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-lp2gx"] Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.542216 5035 scope.go:117] "RemoveContainer" containerID="794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272" Nov 24 19:31:50 crc kubenswrapper[5035]: E1124 19:31:50.542670 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272\": container with ID starting with 794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272 not found: ID does not exist" containerID="794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.542706 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272"} err="failed to get container status \"794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272\": rpc error: code = NotFound desc = could not find container \"794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272\": container with ID starting with 794a3c2e77b89db3b564a49d8da712dd368323639a3fc8b89c9abd4c7977f272 not found: ID does not exist" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.542730 5035 scope.go:117] "RemoveContainer" containerID="8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235" Nov 24 19:31:50 crc kubenswrapper[5035]: E1124 19:31:50.543377 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235\": container with ID starting with 8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235 not found: ID does not exist" containerID="8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235" Nov 24 19:31:50 crc kubenswrapper[5035]: I1124 19:31:50.543575 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235"} err="failed to get container status \"8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235\": rpc error: code = NotFound desc = could not find container \"8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235\": container with ID starting with 8a2b0516400f1184b79fba412ea3c15828b627fef200e31611f8d8f9af047235 not found: ID does not exist" Nov 24 19:31:51 crc kubenswrapper[5035]: I1124 19:31:51.862763 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:51 crc kubenswrapper[5035]: I1124 19:31:51.871981 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.053350 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6sjb\" (UniqueName: \"kubernetes.io/projected/1a40bddb-4969-4100-a311-1fd2614bd9f8-kube-api-access-z6sjb\") pod \"1a40bddb-4969-4100-a311-1fd2614bd9f8\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.053615 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395eef76-f4f1-4543-84b0-0d69abb77653-operator-scripts\") pod \"395eef76-f4f1-4543-84b0-0d69abb77653\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.053658 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb7p4\" (UniqueName: \"kubernetes.io/projected/395eef76-f4f1-4543-84b0-0d69abb77653-kube-api-access-rb7p4\") pod \"395eef76-f4f1-4543-84b0-0d69abb77653\" (UID: \"395eef76-f4f1-4543-84b0-0d69abb77653\") " Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.053708 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a40bddb-4969-4100-a311-1fd2614bd9f8-operator-scripts\") pod \"1a40bddb-4969-4100-a311-1fd2614bd9f8\" (UID: \"1a40bddb-4969-4100-a311-1fd2614bd9f8\") " Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.054638 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a40bddb-4969-4100-a311-1fd2614bd9f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1a40bddb-4969-4100-a311-1fd2614bd9f8" (UID: "1a40bddb-4969-4100-a311-1fd2614bd9f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.054812 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395eef76-f4f1-4543-84b0-0d69abb77653-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "395eef76-f4f1-4543-84b0-0d69abb77653" (UID: "395eef76-f4f1-4543-84b0-0d69abb77653"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.058977 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/395eef76-f4f1-4543-84b0-0d69abb77653-kube-api-access-rb7p4" (OuterVolumeSpecName: "kube-api-access-rb7p4") pod "395eef76-f4f1-4543-84b0-0d69abb77653" (UID: "395eef76-f4f1-4543-84b0-0d69abb77653"). InnerVolumeSpecName "kube-api-access-rb7p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.059396 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a40bddb-4969-4100-a311-1fd2614bd9f8-kube-api-access-z6sjb" (OuterVolumeSpecName: "kube-api-access-z6sjb") pod "1a40bddb-4969-4100-a311-1fd2614bd9f8" (UID: "1a40bddb-4969-4100-a311-1fd2614bd9f8"). InnerVolumeSpecName "kube-api-access-z6sjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.155395 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6sjb\" (UniqueName: \"kubernetes.io/projected/1a40bddb-4969-4100-a311-1fd2614bd9f8-kube-api-access-z6sjb\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.155436 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395eef76-f4f1-4543-84b0-0d69abb77653-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.155461 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb7p4\" (UniqueName: \"kubernetes.io/projected/395eef76-f4f1-4543-84b0-0d69abb77653-kube-api-access-rb7p4\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.155473 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a40bddb-4969-4100-a311-1fd2614bd9f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.218486 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d254b9e-aa95-4296-95aa-52eacc395748" path="/var/lib/kubelet/pods/4d254b9e-aa95-4296-95aa-52eacc395748/volumes" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.424631 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-vhfcj"] Nov 24 19:31:52 crc kubenswrapper[5035]: E1124 19:31:52.425360 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395eef76-f4f1-4543-84b0-0d69abb77653" containerName="mariadb-database-create" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.425396 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="395eef76-f4f1-4543-84b0-0d69abb77653" containerName="mariadb-database-create" Nov 24 19:31:52 crc kubenswrapper[5035]: E1124 19:31:52.425436 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d254b9e-aa95-4296-95aa-52eacc395748" containerName="init" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.425449 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d254b9e-aa95-4296-95aa-52eacc395748" containerName="init" Nov 24 19:31:52 crc kubenswrapper[5035]: E1124 19:31:52.425468 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a40bddb-4969-4100-a311-1fd2614bd9f8" containerName="mariadb-account-create" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.425484 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a40bddb-4969-4100-a311-1fd2614bd9f8" containerName="mariadb-account-create" Nov 24 19:31:52 crc kubenswrapper[5035]: E1124 19:31:52.425506 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d254b9e-aa95-4296-95aa-52eacc395748" containerName="dnsmasq-dns" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.425519 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d254b9e-aa95-4296-95aa-52eacc395748" containerName="dnsmasq-dns" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.425835 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="395eef76-f4f1-4543-84b0-0d69abb77653" containerName="mariadb-database-create" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.425877 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d254b9e-aa95-4296-95aa-52eacc395748" containerName="dnsmasq-dns" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.425921 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a40bddb-4969-4100-a311-1fd2614bd9f8" containerName="mariadb-account-create" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.426955 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.434276 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vhfcj"] Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.513701 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-034b-account-create-g66v8"] Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.514513 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-edc5-account-create-bfvds" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.514719 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-edc5-account-create-bfvds" event={"ID":"1a40bddb-4969-4100-a311-1fd2614bd9f8","Type":"ContainerDied","Data":"5a9a06773df51d39f3d63e8be454521b00e4f0d7e3087054c6982f56d64d3356"} Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.514750 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a9a06773df51d39f3d63e8be454521b00e4f0d7e3087054c6982f56d64d3356" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.514815 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.516282 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmbwf" event={"ID":"395eef76-f4f1-4543-84b0-0d69abb77653","Type":"ContainerDied","Data":"9de59e1249e0ec0c6fdb8cae76ce58cedc0388f766bc0d23d0993808c89355ca"} Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.516342 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9de59e1249e0ec0c6fdb8cae76ce58cedc0388f766bc0d23d0993808c89355ca" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.516404 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmbwf" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.517172 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.527704 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-034b-account-create-g66v8"] Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.562284 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m8pf\" (UniqueName: \"kubernetes.io/projected/c373792a-2d1a-4030-bbb4-3fa4af555411-kube-api-access-5m8pf\") pod \"keystone-db-create-vhfcj\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.562331 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c373792a-2d1a-4030-bbb4-3fa4af555411-operator-scripts\") pod \"keystone-db-create-vhfcj\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.663492 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944131f2-13c8-4e25-b543-bc001f29a85a-operator-scripts\") pod \"keystone-034b-account-create-g66v8\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.663530 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m8pf\" (UniqueName: \"kubernetes.io/projected/c373792a-2d1a-4030-bbb4-3fa4af555411-kube-api-access-5m8pf\") pod \"keystone-db-create-vhfcj\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.663550 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c373792a-2d1a-4030-bbb4-3fa4af555411-operator-scripts\") pod \"keystone-db-create-vhfcj\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.663653 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmznq\" (UniqueName: \"kubernetes.io/projected/944131f2-13c8-4e25-b543-bc001f29a85a-kube-api-access-mmznq\") pod \"keystone-034b-account-create-g66v8\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.664149 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c373792a-2d1a-4030-bbb4-3fa4af555411-operator-scripts\") pod \"keystone-db-create-vhfcj\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.688283 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m8pf\" (UniqueName: \"kubernetes.io/projected/c373792a-2d1a-4030-bbb4-3fa4af555411-kube-api-access-5m8pf\") pod \"keystone-db-create-vhfcj\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.761921 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-z2rqz"] Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.763251 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.763469 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.764616 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmznq\" (UniqueName: \"kubernetes.io/projected/944131f2-13c8-4e25-b543-bc001f29a85a-kube-api-access-mmznq\") pod \"keystone-034b-account-create-g66v8\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.764705 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944131f2-13c8-4e25-b543-bc001f29a85a-operator-scripts\") pod \"keystone-034b-account-create-g66v8\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.765810 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944131f2-13c8-4e25-b543-bc001f29a85a-operator-scripts\") pod \"keystone-034b-account-create-g66v8\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.768549 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z2rqz"] Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.790957 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmznq\" (UniqueName: \"kubernetes.io/projected/944131f2-13c8-4e25-b543-bc001f29a85a-kube-api-access-mmznq\") pod \"keystone-034b-account-create-g66v8\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.838847 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.866207 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/817e8206-e355-4f96-b495-a2afab24f1c4-operator-scripts\") pod \"placement-db-create-z2rqz\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.866338 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76tgb\" (UniqueName: \"kubernetes.io/projected/817e8206-e355-4f96-b495-a2afab24f1c4-kube-api-access-76tgb\") pod \"placement-db-create-z2rqz\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.886420 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a61f-account-create-95bq4"] Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.888027 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.890678 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.905386 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a61f-account-create-95bq4"] Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.968841 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76tgb\" (UniqueName: \"kubernetes.io/projected/817e8206-e355-4f96-b495-a2afab24f1c4-kube-api-access-76tgb\") pod \"placement-db-create-z2rqz\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:52 crc kubenswrapper[5035]: I1124 19:31:52.968919 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/817e8206-e355-4f96-b495-a2afab24f1c4-operator-scripts\") pod \"placement-db-create-z2rqz\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.029432 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/817e8206-e355-4f96-b495-a2afab24f1c4-operator-scripts\") pod \"placement-db-create-z2rqz\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.030110 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76tgb\" (UniqueName: \"kubernetes.io/projected/817e8206-e355-4f96-b495-a2afab24f1c4-kube-api-access-76tgb\") pod \"placement-db-create-z2rqz\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.070410 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-operator-scripts\") pod \"placement-a61f-account-create-95bq4\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.070601 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9vwc\" (UniqueName: \"kubernetes.io/projected/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-kube-api-access-s9vwc\") pod \"placement-a61f-account-create-95bq4\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.172407 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-operator-scripts\") pod \"placement-a61f-account-create-95bq4\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.172579 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9vwc\" (UniqueName: \"kubernetes.io/projected/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-kube-api-access-s9vwc\") pod \"placement-a61f-account-create-95bq4\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.173326 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-operator-scripts\") pod \"placement-a61f-account-create-95bq4\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.189399 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9vwc\" (UniqueName: \"kubernetes.io/projected/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-kube-api-access-s9vwc\") pod \"placement-a61f-account-create-95bq4\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.237475 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vhfcj"] Nov 24 19:31:53 crc kubenswrapper[5035]: W1124 19:31:53.242433 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc373792a_2d1a_4030_bbb4_3fa4af555411.slice/crio-db8756a829b106e3317807c3931dce3acd54eab42c3b99c300791d151280eeca WatchSource:0}: Error finding container db8756a829b106e3317807c3931dce3acd54eab42c3b99c300791d151280eeca: Status 404 returned error can't find the container with id db8756a829b106e3317807c3931dce3acd54eab42c3b99c300791d151280eeca Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.274745 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.330529 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.330701 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-034b-account-create-g66v8"] Nov 24 19:31:53 crc kubenswrapper[5035]: W1124 19:31:53.336093 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod944131f2_13c8_4e25_b543_bc001f29a85a.slice/crio-ffccb4be43701da8f97d3522833c238e61187ebed9e12257f8f5b9042c9c6a82 WatchSource:0}: Error finding container ffccb4be43701da8f97d3522833c238e61187ebed9e12257f8f5b9042c9c6a82: Status 404 returned error can't find the container with id ffccb4be43701da8f97d3522833c238e61187ebed9e12257f8f5b9042c9c6a82 Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.553020 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-034b-account-create-g66v8" event={"ID":"944131f2-13c8-4e25-b543-bc001f29a85a","Type":"ContainerStarted","Data":"cd14a78ee5631624f60504e02515fdfe74264d2f2eb42000a2153830c2f52373"} Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.553411 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-034b-account-create-g66v8" event={"ID":"944131f2-13c8-4e25-b543-bc001f29a85a","Type":"ContainerStarted","Data":"ffccb4be43701da8f97d3522833c238e61187ebed9e12257f8f5b9042c9c6a82"} Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.555250 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vhfcj" event={"ID":"c373792a-2d1a-4030-bbb4-3fa4af555411","Type":"ContainerStarted","Data":"0cb2d0ed629ff5752b787fb0ddb242d5f52d83836f94f0c0ed401ad9638ba9aa"} Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.555495 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vhfcj" event={"ID":"c373792a-2d1a-4030-bbb4-3fa4af555411","Type":"ContainerStarted","Data":"db8756a829b106e3317807c3931dce3acd54eab42c3b99c300791d151280eeca"} Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.555960 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-qrcbh"] Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.557625 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.561037 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-r79bb" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.561758 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.563741 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qrcbh"] Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.570233 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-034b-account-create-g66v8" podStartSLOduration=1.570220694 podStartE2EDuration="1.570220694s" podCreationTimestamp="2025-11-24 19:31:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:31:53.567577242 +0000 UTC m=+1052.090083499" watchObservedRunningTime="2025-11-24 19:31:53.570220694 +0000 UTC m=+1052.092726941" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.658523 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-vhfcj" podStartSLOduration=1.6584989719999998 podStartE2EDuration="1.658498972s" podCreationTimestamp="2025-11-24 19:31:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:31:53.646953735 +0000 UTC m=+1052.169459992" watchObservedRunningTime="2025-11-24 19:31:53.658498972 +0000 UTC m=+1052.181005229" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.708021 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-config-data\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.708374 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-combined-ca-bundle\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.708471 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-db-sync-config-data\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.708500 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56kqh\" (UniqueName: \"kubernetes.io/projected/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-kube-api-access-56kqh\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.733264 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z2rqz"] Nov 24 19:31:53 crc kubenswrapper[5035]: W1124 19:31:53.733774 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod817e8206_e355_4f96_b495_a2afab24f1c4.slice/crio-c90ffa6b22c9fb1e4e433980c148624f48717ab0cbf55130b332e3e4328cd5b5 WatchSource:0}: Error finding container c90ffa6b22c9fb1e4e433980c148624f48717ab0cbf55130b332e3e4328cd5b5: Status 404 returned error can't find the container with id c90ffa6b22c9fb1e4e433980c148624f48717ab0cbf55130b332e3e4328cd5b5 Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.816188 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-db-sync-config-data\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.816238 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56kqh\" (UniqueName: \"kubernetes.io/projected/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-kube-api-access-56kqh\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.816338 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-config-data\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.816360 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-combined-ca-bundle\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.822171 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-db-sync-config-data\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.823671 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-combined-ca-bundle\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.825675 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-config-data\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.850517 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56kqh\" (UniqueName: \"kubernetes.io/projected/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-kube-api-access-56kqh\") pod \"glance-db-sync-qrcbh\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.853758 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a61f-account-create-95bq4"] Nov 24 19:31:53 crc kubenswrapper[5035]: I1124 19:31:53.878202 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qrcbh" Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.409749 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qrcbh"] Nov 24 19:31:54 crc kubenswrapper[5035]: W1124 19:31:54.490486 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod030eac16_e8f0_4a3c_9d1e_588cc25d6ebe.slice/crio-ebdfe74cc37e9c81bea6d7d08b3b1d5efe8eceefe93cc2ab0299cfc91f59b9d9 WatchSource:0}: Error finding container ebdfe74cc37e9c81bea6d7d08b3b1d5efe8eceefe93cc2ab0299cfc91f59b9d9: Status 404 returned error can't find the container with id ebdfe74cc37e9c81bea6d7d08b3b1d5efe8eceefe93cc2ab0299cfc91f59b9d9 Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.563597 5035 generic.go:334] "Generic (PLEG): container finished" podID="817e8206-e355-4f96-b495-a2afab24f1c4" containerID="2ca0cee62ad3f7a297bfcda9bc92aef39dfd13f7ec4bc0c6dbe1f9d0d6ca24be" exitCode=0 Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.563651 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z2rqz" event={"ID":"817e8206-e355-4f96-b495-a2afab24f1c4","Type":"ContainerDied","Data":"2ca0cee62ad3f7a297bfcda9bc92aef39dfd13f7ec4bc0c6dbe1f9d0d6ca24be"} Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.564020 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z2rqz" event={"ID":"817e8206-e355-4f96-b495-a2afab24f1c4","Type":"ContainerStarted","Data":"c90ffa6b22c9fb1e4e433980c148624f48717ab0cbf55130b332e3e4328cd5b5"} Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.565583 5035 generic.go:334] "Generic (PLEG): container finished" podID="e8f3e40f-fd65-4c1c-b3e9-7d2158a97038" containerID="f54a44ccc3e2b2712001515b65b9420624b3c71fbf91f2786740003a91fd887e" exitCode=0 Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.565652 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a61f-account-create-95bq4" event={"ID":"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038","Type":"ContainerDied","Data":"f54a44ccc3e2b2712001515b65b9420624b3c71fbf91f2786740003a91fd887e"} Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.565685 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a61f-account-create-95bq4" event={"ID":"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038","Type":"ContainerStarted","Data":"19a9954950dfd5f392c765cf61313c541cbc683fb78b2020cc2cb0bff52ddddc"} Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.567393 5035 generic.go:334] "Generic (PLEG): container finished" podID="c373792a-2d1a-4030-bbb4-3fa4af555411" containerID="0cb2d0ed629ff5752b787fb0ddb242d5f52d83836f94f0c0ed401ad9638ba9aa" exitCode=0 Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.567457 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vhfcj" event={"ID":"c373792a-2d1a-4030-bbb4-3fa4af555411","Type":"ContainerDied","Data":"0cb2d0ed629ff5752b787fb0ddb242d5f52d83836f94f0c0ed401ad9638ba9aa"} Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.569007 5035 generic.go:334] "Generic (PLEG): container finished" podID="944131f2-13c8-4e25-b543-bc001f29a85a" containerID="cd14a78ee5631624f60504e02515fdfe74264d2f2eb42000a2153830c2f52373" exitCode=0 Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.569072 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-034b-account-create-g66v8" event={"ID":"944131f2-13c8-4e25-b543-bc001f29a85a","Type":"ContainerDied","Data":"cd14a78ee5631624f60504e02515fdfe74264d2f2eb42000a2153830c2f52373"} Nov 24 19:31:54 crc kubenswrapper[5035]: I1124 19:31:54.569979 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qrcbh" event={"ID":"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe","Type":"ContainerStarted","Data":"ebdfe74cc37e9c81bea6d7d08b3b1d5efe8eceefe93cc2ab0299cfc91f59b9d9"} Nov 24 19:31:55 crc kubenswrapper[5035]: I1124 19:31:55.948400 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:55 crc kubenswrapper[5035]: I1124 19:31:55.950172 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmznq\" (UniqueName: \"kubernetes.io/projected/944131f2-13c8-4e25-b543-bc001f29a85a-kube-api-access-mmznq\") pod \"944131f2-13c8-4e25-b543-bc001f29a85a\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " Nov 24 19:31:55 crc kubenswrapper[5035]: I1124 19:31:55.950268 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944131f2-13c8-4e25-b543-bc001f29a85a-operator-scripts\") pod \"944131f2-13c8-4e25-b543-bc001f29a85a\" (UID: \"944131f2-13c8-4e25-b543-bc001f29a85a\") " Nov 24 19:31:55 crc kubenswrapper[5035]: I1124 19:31:55.951534 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/944131f2-13c8-4e25-b543-bc001f29a85a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "944131f2-13c8-4e25-b543-bc001f29a85a" (UID: "944131f2-13c8-4e25-b543-bc001f29a85a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:55 crc kubenswrapper[5035]: I1124 19:31:55.960539 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944131f2-13c8-4e25-b543-bc001f29a85a-kube-api-access-mmznq" (OuterVolumeSpecName: "kube-api-access-mmznq") pod "944131f2-13c8-4e25-b543-bc001f29a85a" (UID: "944131f2-13c8-4e25-b543-bc001f29a85a"). InnerVolumeSpecName "kube-api-access-mmznq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.031676 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.036356 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.048809 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.051171 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/817e8206-e355-4f96-b495-a2afab24f1c4-operator-scripts\") pod \"817e8206-e355-4f96-b495-a2afab24f1c4\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.051357 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5m8pf\" (UniqueName: \"kubernetes.io/projected/c373792a-2d1a-4030-bbb4-3fa4af555411-kube-api-access-5m8pf\") pod \"c373792a-2d1a-4030-bbb4-3fa4af555411\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.051650 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76tgb\" (UniqueName: \"kubernetes.io/projected/817e8206-e355-4f96-b495-a2afab24f1c4-kube-api-access-76tgb\") pod \"817e8206-e355-4f96-b495-a2afab24f1c4\" (UID: \"817e8206-e355-4f96-b495-a2afab24f1c4\") " Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.051702 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9vwc\" (UniqueName: \"kubernetes.io/projected/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-kube-api-access-s9vwc\") pod \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.051755 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c373792a-2d1a-4030-bbb4-3fa4af555411-operator-scripts\") pod \"c373792a-2d1a-4030-bbb4-3fa4af555411\" (UID: \"c373792a-2d1a-4030-bbb4-3fa4af555411\") " Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.051806 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-operator-scripts\") pod \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\" (UID: \"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038\") " Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.052352 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/944131f2-13c8-4e25-b543-bc001f29a85a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.052373 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmznq\" (UniqueName: \"kubernetes.io/projected/944131f2-13c8-4e25-b543-bc001f29a85a-kube-api-access-mmznq\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.052773 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8f3e40f-fd65-4c1c-b3e9-7d2158a97038" (UID: "e8f3e40f-fd65-4c1c-b3e9-7d2158a97038"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.054345 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c373792a-2d1a-4030-bbb4-3fa4af555411-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c373792a-2d1a-4030-bbb4-3fa4af555411" (UID: "c373792a-2d1a-4030-bbb4-3fa4af555411"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.054473 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/817e8206-e355-4f96-b495-a2afab24f1c4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "817e8206-e355-4f96-b495-a2afab24f1c4" (UID: "817e8206-e355-4f96-b495-a2afab24f1c4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.058167 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c373792a-2d1a-4030-bbb4-3fa4af555411-kube-api-access-5m8pf" (OuterVolumeSpecName: "kube-api-access-5m8pf") pod "c373792a-2d1a-4030-bbb4-3fa4af555411" (UID: "c373792a-2d1a-4030-bbb4-3fa4af555411"). InnerVolumeSpecName "kube-api-access-5m8pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.062408 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/817e8206-e355-4f96-b495-a2afab24f1c4-kube-api-access-76tgb" (OuterVolumeSpecName: "kube-api-access-76tgb") pod "817e8206-e355-4f96-b495-a2afab24f1c4" (UID: "817e8206-e355-4f96-b495-a2afab24f1c4"). InnerVolumeSpecName "kube-api-access-76tgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.067484 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-kube-api-access-s9vwc" (OuterVolumeSpecName: "kube-api-access-s9vwc") pod "e8f3e40f-fd65-4c1c-b3e9-7d2158a97038" (UID: "e8f3e40f-fd65-4c1c-b3e9-7d2158a97038"). InnerVolumeSpecName "kube-api-access-s9vwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.153531 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/817e8206-e355-4f96-b495-a2afab24f1c4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.153567 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5m8pf\" (UniqueName: \"kubernetes.io/projected/c373792a-2d1a-4030-bbb4-3fa4af555411-kube-api-access-5m8pf\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.153582 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76tgb\" (UniqueName: \"kubernetes.io/projected/817e8206-e355-4f96-b495-a2afab24f1c4-kube-api-access-76tgb\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.153594 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9vwc\" (UniqueName: \"kubernetes.io/projected/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-kube-api-access-s9vwc\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.153606 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c373792a-2d1a-4030-bbb4-3fa4af555411-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.153619 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:31:56 crc kubenswrapper[5035]: E1124 19:31:56.398258 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8f3e40f_fd65_4c1c_b3e9_7d2158a97038.slice/crio-19a9954950dfd5f392c765cf61313c541cbc683fb78b2020cc2cb0bff52ddddc\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod944131f2_13c8_4e25_b543_bc001f29a85a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8f3e40f_fd65_4c1c_b3e9_7d2158a97038.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod817e8206_e355_4f96_b495_a2afab24f1c4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc373792a_2d1a_4030_bbb4_3fa4af555411.slice\": RecentStats: unable to find data in memory cache]" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.595888 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z2rqz" event={"ID":"817e8206-e355-4f96-b495-a2afab24f1c4","Type":"ContainerDied","Data":"c90ffa6b22c9fb1e4e433980c148624f48717ab0cbf55130b332e3e4328cd5b5"} Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.595941 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c90ffa6b22c9fb1e4e433980c148624f48717ab0cbf55130b332e3e4328cd5b5" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.595907 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z2rqz" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.599302 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a61f-account-create-95bq4" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.599737 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a61f-account-create-95bq4" event={"ID":"e8f3e40f-fd65-4c1c-b3e9-7d2158a97038","Type":"ContainerDied","Data":"19a9954950dfd5f392c765cf61313c541cbc683fb78b2020cc2cb0bff52ddddc"} Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.599775 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19a9954950dfd5f392c765cf61313c541cbc683fb78b2020cc2cb0bff52ddddc" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.601844 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vhfcj" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.601848 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vhfcj" event={"ID":"c373792a-2d1a-4030-bbb4-3fa4af555411","Type":"ContainerDied","Data":"db8756a829b106e3317807c3931dce3acd54eab42c3b99c300791d151280eeca"} Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.601898 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db8756a829b106e3317807c3931dce3acd54eab42c3b99c300791d151280eeca" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.603200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-034b-account-create-g66v8" event={"ID":"944131f2-13c8-4e25-b543-bc001f29a85a","Type":"ContainerDied","Data":"ffccb4be43701da8f97d3522833c238e61187ebed9e12257f8f5b9042c9c6a82"} Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.603221 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffccb4be43701da8f97d3522833c238e61187ebed9e12257f8f5b9042c9c6a82" Nov 24 19:31:56 crc kubenswrapper[5035]: I1124 19:31:56.603324 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-034b-account-create-g66v8" Nov 24 19:31:58 crc kubenswrapper[5035]: I1124 19:31:58.618317 5035 generic.go:334] "Generic (PLEG): container finished" podID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerID="13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04" exitCode=0 Nov 24 19:31:58 crc kubenswrapper[5035]: I1124 19:31:58.618373 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ed0e7cb-164f-4689-8714-1f11bfa25725","Type":"ContainerDied","Data":"13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04"} Nov 24 19:31:58 crc kubenswrapper[5035]: I1124 19:31:58.620454 5035 generic.go:334] "Generic (PLEG): container finished" podID="8716e4c9-35b0-4922-afc2-13006d78283a" containerID="821d307dfb370f447c67d71297fcefc4e81c98065f7a613d6310b10ef3ad01a6" exitCode=0 Nov 24 19:31:58 crc kubenswrapper[5035]: I1124 19:31:58.620473 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8716e4c9-35b0-4922-afc2-13006d78283a","Type":"ContainerDied","Data":"821d307dfb370f447c67d71297fcefc4e81c98065f7a613d6310b10ef3ad01a6"} Nov 24 19:31:59 crc kubenswrapper[5035]: I1124 19:31:59.636783 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ed0e7cb-164f-4689-8714-1f11bfa25725","Type":"ContainerStarted","Data":"614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f"} Nov 24 19:31:59 crc kubenswrapper[5035]: I1124 19:31:59.637690 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 19:31:59 crc kubenswrapper[5035]: I1124 19:31:59.639751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8716e4c9-35b0-4922-afc2-13006d78283a","Type":"ContainerStarted","Data":"4977ad3c8b82328502e1b7cffa7822db6d743aec4c461b5644175f59a455814d"} Nov 24 19:31:59 crc kubenswrapper[5035]: I1124 19:31:59.639989 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:31:59 crc kubenswrapper[5035]: I1124 19:31:59.667018 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.930595722 podStartE2EDuration="51.667000851s" podCreationTimestamp="2025-11-24 19:31:08 +0000 UTC" firstStartedPulling="2025-11-24 19:31:10.027503249 +0000 UTC m=+1008.550009496" lastFinishedPulling="2025-11-24 19:31:24.763908368 +0000 UTC m=+1023.286414625" observedRunningTime="2025-11-24 19:31:59.664954365 +0000 UTC m=+1058.187460692" watchObservedRunningTime="2025-11-24 19:31:59.667000851 +0000 UTC m=+1058.189507098" Nov 24 19:31:59 crc kubenswrapper[5035]: I1124 19:31:59.705279 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.338275156 podStartE2EDuration="51.705252708s" podCreationTimestamp="2025-11-24 19:31:08 +0000 UTC" firstStartedPulling="2025-11-24 19:31:10.424628174 +0000 UTC m=+1008.947134421" lastFinishedPulling="2025-11-24 19:31:24.791605716 +0000 UTC m=+1023.314111973" observedRunningTime="2025-11-24 19:31:59.697397713 +0000 UTC m=+1058.219903980" watchObservedRunningTime="2025-11-24 19:31:59.705252708 +0000 UTC m=+1058.227758995" Nov 24 19:32:01 crc kubenswrapper[5035]: I1124 19:32:01.089878 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.658764 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-82r49" podUID="aa0e4fdb-6066-4163-9864-9e52443afdb0" containerName="ovn-controller" probeResult="failure" output=< Nov 24 19:32:04 crc kubenswrapper[5035]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 19:32:04 crc kubenswrapper[5035]: > Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.696849 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.702080 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6l2z4" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.916762 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-82r49-config-9g2z6"] Nov 24 19:32:04 crc kubenswrapper[5035]: E1124 19:32:04.917346 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f3e40f-fd65-4c1c-b3e9-7d2158a97038" containerName="mariadb-account-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.917433 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f3e40f-fd65-4c1c-b3e9-7d2158a97038" containerName="mariadb-account-create" Nov 24 19:32:04 crc kubenswrapper[5035]: E1124 19:32:04.917497 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c373792a-2d1a-4030-bbb4-3fa4af555411" containerName="mariadb-database-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.917553 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c373792a-2d1a-4030-bbb4-3fa4af555411" containerName="mariadb-database-create" Nov 24 19:32:04 crc kubenswrapper[5035]: E1124 19:32:04.917618 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="817e8206-e355-4f96-b495-a2afab24f1c4" containerName="mariadb-database-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.917668 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="817e8206-e355-4f96-b495-a2afab24f1c4" containerName="mariadb-database-create" Nov 24 19:32:04 crc kubenswrapper[5035]: E1124 19:32:04.917733 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944131f2-13c8-4e25-b543-bc001f29a85a" containerName="mariadb-account-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.917799 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="944131f2-13c8-4e25-b543-bc001f29a85a" containerName="mariadb-account-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.917998 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="817e8206-e355-4f96-b495-a2afab24f1c4" containerName="mariadb-database-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.918061 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="944131f2-13c8-4e25-b543-bc001f29a85a" containerName="mariadb-account-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.918129 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c373792a-2d1a-4030-bbb4-3fa4af555411" containerName="mariadb-database-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.918183 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8f3e40f-fd65-4c1c-b3e9-7d2158a97038" containerName="mariadb-account-create" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.919089 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.921318 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 19:32:04 crc kubenswrapper[5035]: I1124 19:32:04.932986 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-82r49-config-9g2z6"] Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.005739 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-additional-scripts\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.005818 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-log-ovn\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.005906 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.006281 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kqzt\" (UniqueName: \"kubernetes.io/projected/2f4a70fa-2496-4b19-9f39-539078bec5c9-kube-api-access-4kqzt\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.006369 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run-ovn\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.006459 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-scripts\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.107538 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-scripts\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.107612 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-additional-scripts\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.107657 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-log-ovn\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.107674 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.107723 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kqzt\" (UniqueName: \"kubernetes.io/projected/2f4a70fa-2496-4b19-9f39-539078bec5c9-kube-api-access-4kqzt\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.107747 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run-ovn\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.107997 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run-ovn\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.108038 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.108160 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-log-ovn\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.108430 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-additional-scripts\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.109701 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-scripts\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.128445 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kqzt\" (UniqueName: \"kubernetes.io/projected/2f4a70fa-2496-4b19-9f39-539078bec5c9-kube-api-access-4kqzt\") pod \"ovn-controller-82r49-config-9g2z6\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:05 crc kubenswrapper[5035]: I1124 19:32:05.242122 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:06 crc kubenswrapper[5035]: I1124 19:32:06.897915 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-82r49-config-9g2z6"] Nov 24 19:32:06 crc kubenswrapper[5035]: W1124 19:32:06.907157 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4a70fa_2496_4b19_9f39_539078bec5c9.slice/crio-f9bc327c253c489a9a9d03ca3ca201fe24fadfe07b6f95102b8816e8271acf79 WatchSource:0}: Error finding container f9bc327c253c489a9a9d03ca3ca201fe24fadfe07b6f95102b8816e8271acf79: Status 404 returned error can't find the container with id f9bc327c253c489a9a9d03ca3ca201fe24fadfe07b6f95102b8816e8271acf79 Nov 24 19:32:07 crc kubenswrapper[5035]: I1124 19:32:07.719411 5035 generic.go:334] "Generic (PLEG): container finished" podID="2f4a70fa-2496-4b19-9f39-539078bec5c9" containerID="c39bf0115ddd19cdb161ec48d1442aa37d41032ad0df9379f64a096d3cdded02" exitCode=0 Nov 24 19:32:07 crc kubenswrapper[5035]: I1124 19:32:07.719553 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-82r49-config-9g2z6" event={"ID":"2f4a70fa-2496-4b19-9f39-539078bec5c9","Type":"ContainerDied","Data":"c39bf0115ddd19cdb161ec48d1442aa37d41032ad0df9379f64a096d3cdded02"} Nov 24 19:32:07 crc kubenswrapper[5035]: I1124 19:32:07.719795 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-82r49-config-9g2z6" event={"ID":"2f4a70fa-2496-4b19-9f39-539078bec5c9","Type":"ContainerStarted","Data":"f9bc327c253c489a9a9d03ca3ca201fe24fadfe07b6f95102b8816e8271acf79"} Nov 24 19:32:07 crc kubenswrapper[5035]: I1124 19:32:07.732256 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qrcbh" event={"ID":"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe","Type":"ContainerStarted","Data":"4cce725c77ece9593b000a253ae81b3eb7df00c2a24ec012d0b1f9d34272c134"} Nov 24 19:32:07 crc kubenswrapper[5035]: I1124 19:32:07.780612 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-qrcbh" podStartSLOduration=2.69788803 podStartE2EDuration="14.780578907s" podCreationTimestamp="2025-11-24 19:31:53 +0000 UTC" firstStartedPulling="2025-11-24 19:31:54.493106927 +0000 UTC m=+1053.015613184" lastFinishedPulling="2025-11-24 19:32:06.575797804 +0000 UTC m=+1065.098304061" observedRunningTime="2025-11-24 19:32:07.769287628 +0000 UTC m=+1066.291793935" watchObservedRunningTime="2025-11-24 19:32:07.780578907 +0000 UTC m=+1066.303085194" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.025390 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.181477 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kqzt\" (UniqueName: \"kubernetes.io/projected/2f4a70fa-2496-4b19-9f39-539078bec5c9-kube-api-access-4kqzt\") pod \"2f4a70fa-2496-4b19-9f39-539078bec5c9\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.181552 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-log-ovn\") pod \"2f4a70fa-2496-4b19-9f39-539078bec5c9\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.181600 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run\") pod \"2f4a70fa-2496-4b19-9f39-539078bec5c9\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.181687 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "2f4a70fa-2496-4b19-9f39-539078bec5c9" (UID: "2f4a70fa-2496-4b19-9f39-539078bec5c9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.181644 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-additional-scripts\") pod \"2f4a70fa-2496-4b19-9f39-539078bec5c9\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.181701 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run" (OuterVolumeSpecName: "var-run") pod "2f4a70fa-2496-4b19-9f39-539078bec5c9" (UID: "2f4a70fa-2496-4b19-9f39-539078bec5c9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.181980 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run-ovn\") pod \"2f4a70fa-2496-4b19-9f39-539078bec5c9\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.182037 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "2f4a70fa-2496-4b19-9f39-539078bec5c9" (UID: "2f4a70fa-2496-4b19-9f39-539078bec5c9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.182109 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-scripts\") pod \"2f4a70fa-2496-4b19-9f39-539078bec5c9\" (UID: \"2f4a70fa-2496-4b19-9f39-539078bec5c9\") " Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.182475 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "2f4a70fa-2496-4b19-9f39-539078bec5c9" (UID: "2f4a70fa-2496-4b19-9f39-539078bec5c9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.182886 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-scripts" (OuterVolumeSpecName: "scripts") pod "2f4a70fa-2496-4b19-9f39-539078bec5c9" (UID: "2f4a70fa-2496-4b19-9f39-539078bec5c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.183448 5035 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.183465 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.183473 5035 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.183481 5035 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2f4a70fa-2496-4b19-9f39-539078bec5c9-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.183491 5035 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2f4a70fa-2496-4b19-9f39-539078bec5c9-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.193475 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4a70fa-2496-4b19-9f39-539078bec5c9-kube-api-access-4kqzt" (OuterVolumeSpecName: "kube-api-access-4kqzt") pod "2f4a70fa-2496-4b19-9f39-539078bec5c9" (UID: "2f4a70fa-2496-4b19-9f39-539078bec5c9"). InnerVolumeSpecName "kube-api-access-4kqzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.285612 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kqzt\" (UniqueName: \"kubernetes.io/projected/2f4a70fa-2496-4b19-9f39-539078bec5c9-kube-api-access-4kqzt\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.517629 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.687946 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-82r49" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.748163 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-82r49-config-9g2z6" event={"ID":"2f4a70fa-2496-4b19-9f39-539078bec5c9","Type":"ContainerDied","Data":"f9bc327c253c489a9a9d03ca3ca201fe24fadfe07b6f95102b8816e8271acf79"} Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.748199 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9bc327c253c489a9a9d03ca3ca201fe24fadfe07b6f95102b8816e8271acf79" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.748480 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-82r49-config-9g2z6" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.830520 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dgz8h"] Nov 24 19:32:09 crc kubenswrapper[5035]: E1124 19:32:09.831081 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4a70fa-2496-4b19-9f39-539078bec5c9" containerName="ovn-config" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.831256 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4a70fa-2496-4b19-9f39-539078bec5c9" containerName="ovn-config" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.831516 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4a70fa-2496-4b19-9f39-539078bec5c9" containerName="ovn-config" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.832048 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.842735 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dgz8h"] Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.882743 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.895937 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68fcf962-da19-4161-8eee-aecde6615ba4-operator-scripts\") pod \"barbican-db-create-dgz8h\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.895989 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x64hg\" (UniqueName: \"kubernetes.io/projected/68fcf962-da19-4161-8eee-aecde6615ba4-kube-api-access-x64hg\") pod \"barbican-db-create-dgz8h\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.924902 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d325-account-create-dwc79"] Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.925919 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.930427 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d325-account-create-dwc79"] Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.932608 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.997563 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d84b9e7-688c-4926-bdce-1581f89d4875-operator-scripts\") pod \"barbican-d325-account-create-dwc79\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.997642 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx99t\" (UniqueName: \"kubernetes.io/projected/8d84b9e7-688c-4926-bdce-1581f89d4875-kube-api-access-zx99t\") pod \"barbican-d325-account-create-dwc79\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.997982 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68fcf962-da19-4161-8eee-aecde6615ba4-operator-scripts\") pod \"barbican-db-create-dgz8h\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.998069 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x64hg\" (UniqueName: \"kubernetes.io/projected/68fcf962-da19-4161-8eee-aecde6615ba4-kube-api-access-x64hg\") pod \"barbican-db-create-dgz8h\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:09 crc kubenswrapper[5035]: I1124 19:32:09.998842 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68fcf962-da19-4161-8eee-aecde6615ba4-operator-scripts\") pod \"barbican-db-create-dgz8h\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.017177 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vrrsl"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.018238 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.021884 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x64hg\" (UniqueName: \"kubernetes.io/projected/68fcf962-da19-4161-8eee-aecde6615ba4-kube-api-access-x64hg\") pod \"barbican-db-create-dgz8h\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.027144 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c419-account-create-nxsz2"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.028109 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.029546 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.054045 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c419-account-create-nxsz2"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.087824 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vrrsl"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.109381 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d84b9e7-688c-4926-bdce-1581f89d4875-operator-scripts\") pod \"barbican-d325-account-create-dwc79\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.109436 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e080c702-5f10-4d8a-a192-07466d0d7644-operator-scripts\") pod \"cinder-db-create-vrrsl\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.109473 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btg56\" (UniqueName: \"kubernetes.io/projected/e080c702-5f10-4d8a-a192-07466d0d7644-kube-api-access-btg56\") pod \"cinder-db-create-vrrsl\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.109500 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb565\" (UniqueName: \"kubernetes.io/projected/80412713-2ef8-4499-ae54-2c6aebac5377-kube-api-access-zb565\") pod \"cinder-c419-account-create-nxsz2\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.109604 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80412713-2ef8-4499-ae54-2c6aebac5377-operator-scripts\") pod \"cinder-c419-account-create-nxsz2\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.109679 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx99t\" (UniqueName: \"kubernetes.io/projected/8d84b9e7-688c-4926-bdce-1581f89d4875-kube-api-access-zx99t\") pod \"barbican-d325-account-create-dwc79\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.110108 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d84b9e7-688c-4926-bdce-1581f89d4875-operator-scripts\") pod \"barbican-d325-account-create-dwc79\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.130872 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx99t\" (UniqueName: \"kubernetes.io/projected/8d84b9e7-688c-4926-bdce-1581f89d4875-kube-api-access-zx99t\") pod \"barbican-d325-account-create-dwc79\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.147977 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.167981 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-82r49-config-9g2z6"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.177192 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-82r49-config-9g2z6"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.210710 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e080c702-5f10-4d8a-a192-07466d0d7644-operator-scripts\") pod \"cinder-db-create-vrrsl\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.210752 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btg56\" (UniqueName: \"kubernetes.io/projected/e080c702-5f10-4d8a-a192-07466d0d7644-kube-api-access-btg56\") pod \"cinder-db-create-vrrsl\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.210771 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb565\" (UniqueName: \"kubernetes.io/projected/80412713-2ef8-4499-ae54-2c6aebac5377-kube-api-access-zb565\") pod \"cinder-c419-account-create-nxsz2\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.210797 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80412713-2ef8-4499-ae54-2c6aebac5377-operator-scripts\") pod \"cinder-c419-account-create-nxsz2\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.211518 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e080c702-5f10-4d8a-a192-07466d0d7644-operator-scripts\") pod \"cinder-db-create-vrrsl\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.211793 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80412713-2ef8-4499-ae54-2c6aebac5377-operator-scripts\") pod \"cinder-c419-account-create-nxsz2\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.224610 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f4a70fa-2496-4b19-9f39-539078bec5c9" path="/var/lib/kubelet/pods/2f4a70fa-2496-4b19-9f39-539078bec5c9/volumes" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.237805 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb565\" (UniqueName: \"kubernetes.io/projected/80412713-2ef8-4499-ae54-2c6aebac5377-kube-api-access-zb565\") pod \"cinder-c419-account-create-nxsz2\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.246063 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.268012 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btg56\" (UniqueName: \"kubernetes.io/projected/e080c702-5f10-4d8a-a192-07466d0d7644-kube-api-access-btg56\") pod \"cinder-db-create-vrrsl\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.288102 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-9jwdn"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.294234 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.298326 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2trk2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.298466 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.298514 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.298725 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.313763 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9jwdn"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.333928 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-cxbtl"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.334941 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.355731 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f85b-account-create-zpj9j"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.356751 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.363692 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.373397 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cxbtl"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.382945 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.387209 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f85b-account-create-zpj9j"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.392161 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.437902 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4gmr\" (UniqueName: \"kubernetes.io/projected/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-kube-api-access-j4gmr\") pod \"neutron-f85b-account-create-zpj9j\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.437950 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qdvg\" (UniqueName: \"kubernetes.io/projected/7ba53ebf-bae3-4629-af0b-8637e6491515-kube-api-access-2qdvg\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.437973 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-operator-scripts\") pod \"neutron-f85b-account-create-zpj9j\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.438002 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqwnp\" (UniqueName: \"kubernetes.io/projected/da1e9bbe-02f8-4eda-be3d-bf882329f44c-kube-api-access-qqwnp\") pod \"neutron-db-create-cxbtl\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.438020 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-config-data\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.438042 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da1e9bbe-02f8-4eda-be3d-bf882329f44c-operator-scripts\") pod \"neutron-db-create-cxbtl\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.438112 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-combined-ca-bundle\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.539963 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-combined-ca-bundle\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.540268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4gmr\" (UniqueName: \"kubernetes.io/projected/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-kube-api-access-j4gmr\") pod \"neutron-f85b-account-create-zpj9j\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.540486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qdvg\" (UniqueName: \"kubernetes.io/projected/7ba53ebf-bae3-4629-af0b-8637e6491515-kube-api-access-2qdvg\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.540515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-operator-scripts\") pod \"neutron-f85b-account-create-zpj9j\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.540547 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqwnp\" (UniqueName: \"kubernetes.io/projected/da1e9bbe-02f8-4eda-be3d-bf882329f44c-kube-api-access-qqwnp\") pod \"neutron-db-create-cxbtl\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.540565 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-config-data\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.540585 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da1e9bbe-02f8-4eda-be3d-bf882329f44c-operator-scripts\") pod \"neutron-db-create-cxbtl\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.541350 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da1e9bbe-02f8-4eda-be3d-bf882329f44c-operator-scripts\") pod \"neutron-db-create-cxbtl\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.541353 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-operator-scripts\") pod \"neutron-f85b-account-create-zpj9j\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.546486 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-combined-ca-bundle\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.550903 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-config-data\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.561434 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qdvg\" (UniqueName: \"kubernetes.io/projected/7ba53ebf-bae3-4629-af0b-8637e6491515-kube-api-access-2qdvg\") pod \"keystone-db-sync-9jwdn\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.562989 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqwnp\" (UniqueName: \"kubernetes.io/projected/da1e9bbe-02f8-4eda-be3d-bf882329f44c-kube-api-access-qqwnp\") pod \"neutron-db-create-cxbtl\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.563002 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4gmr\" (UniqueName: \"kubernetes.io/projected/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-kube-api-access-j4gmr\") pod \"neutron-f85b-account-create-zpj9j\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.612830 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.659845 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.678649 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.839056 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d325-account-create-dwc79"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.915461 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dgz8h"] Nov 24 19:32:10 crc kubenswrapper[5035]: I1124 19:32:10.988364 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c419-account-create-nxsz2"] Nov 24 19:32:10 crc kubenswrapper[5035]: W1124 19:32:10.999860 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode080c702_5f10_4d8a_a192_07466d0d7644.slice/crio-4e9f9be866f32e784880811da2b4a009fcc952108c9760ed8ee31c4f4f7942c0 WatchSource:0}: Error finding container 4e9f9be866f32e784880811da2b4a009fcc952108c9760ed8ee31c4f4f7942c0: Status 404 returned error can't find the container with id 4e9f9be866f32e784880811da2b4a009fcc952108c9760ed8ee31c4f4f7942c0 Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.015265 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vrrsl"] Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.116641 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9jwdn"] Nov 24 19:32:11 crc kubenswrapper[5035]: W1124 19:32:11.152173 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ba53ebf_bae3_4629_af0b_8637e6491515.slice/crio-70bcb384ad3f7471aac600c64360ffd39116654b4ec3fda042e1e532de1c0d5c WatchSource:0}: Error finding container 70bcb384ad3f7471aac600c64360ffd39116654b4ec3fda042e1e532de1c0d5c: Status 404 returned error can't find the container with id 70bcb384ad3f7471aac600c64360ffd39116654b4ec3fda042e1e532de1c0d5c Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.204117 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f85b-account-create-zpj9j"] Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.212674 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cxbtl"] Nov 24 19:32:11 crc kubenswrapper[5035]: W1124 19:32:11.217326 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaeff8377_e685_4e82_ad78_b9b99f9e1bc0.slice/crio-9952a67c048da2ff42b23962a89a63e15d4c12c1112e67ad64a7b8eb98f3c7aa WatchSource:0}: Error finding container 9952a67c048da2ff42b23962a89a63e15d4c12c1112e67ad64a7b8eb98f3c7aa: Status 404 returned error can't find the container with id 9952a67c048da2ff42b23962a89a63e15d4c12c1112e67ad64a7b8eb98f3c7aa Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.768650 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cxbtl" event={"ID":"da1e9bbe-02f8-4eda-be3d-bf882329f44c","Type":"ContainerStarted","Data":"2b65dff80626f7d183a900f4cf6dc685afc94965cc5638d12d4419aa23d9d680"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.768699 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cxbtl" event={"ID":"da1e9bbe-02f8-4eda-be3d-bf882329f44c","Type":"ContainerStarted","Data":"7e03111bb7eb58f5066cd291710d626ce3ef6206fca58440033368c0b96275a1"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.770651 5035 generic.go:334] "Generic (PLEG): container finished" podID="e080c702-5f10-4d8a-a192-07466d0d7644" containerID="492f865652f7fcd134fb8860f0d4568381a96e467551d261b344b91454c4f9f9" exitCode=0 Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.770707 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vrrsl" event={"ID":"e080c702-5f10-4d8a-a192-07466d0d7644","Type":"ContainerDied","Data":"492f865652f7fcd134fb8860f0d4568381a96e467551d261b344b91454c4f9f9"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.770740 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vrrsl" event={"ID":"e080c702-5f10-4d8a-a192-07466d0d7644","Type":"ContainerStarted","Data":"4e9f9be866f32e784880811da2b4a009fcc952108c9760ed8ee31c4f4f7942c0"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.772433 5035 generic.go:334] "Generic (PLEG): container finished" podID="68fcf962-da19-4161-8eee-aecde6615ba4" containerID="8521c323661c784a4292e298b4f92f72438e6d82650f8d6785f444a848ff7881" exitCode=0 Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.772468 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dgz8h" event={"ID":"68fcf962-da19-4161-8eee-aecde6615ba4","Type":"ContainerDied","Data":"8521c323661c784a4292e298b4f92f72438e6d82650f8d6785f444a848ff7881"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.772495 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dgz8h" event={"ID":"68fcf962-da19-4161-8eee-aecde6615ba4","Type":"ContainerStarted","Data":"07b0f55bef9eef64887643db4f6540ace9090e080678b1f78c338a0bd5c69006"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.774554 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f85b-account-create-zpj9j" event={"ID":"aeff8377-e685-4e82-ad78-b9b99f9e1bc0","Type":"ContainerStarted","Data":"805534dd8880ebe48abf9b3a13b871ab38cd236f5d2fb19d4e1a9388254c77d9"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.774614 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f85b-account-create-zpj9j" event={"ID":"aeff8377-e685-4e82-ad78-b9b99f9e1bc0","Type":"ContainerStarted","Data":"9952a67c048da2ff42b23962a89a63e15d4c12c1112e67ad64a7b8eb98f3c7aa"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.776895 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9jwdn" event={"ID":"7ba53ebf-bae3-4629-af0b-8637e6491515","Type":"ContainerStarted","Data":"70bcb384ad3f7471aac600c64360ffd39116654b4ec3fda042e1e532de1c0d5c"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.779727 5035 generic.go:334] "Generic (PLEG): container finished" podID="80412713-2ef8-4499-ae54-2c6aebac5377" containerID="d8da86f1d1cd555086611c05df2b0b0c4b17b03065becb8d4fcd1bf9d2b66b32" exitCode=0 Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.779855 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c419-account-create-nxsz2" event={"ID":"80412713-2ef8-4499-ae54-2c6aebac5377","Type":"ContainerDied","Data":"d8da86f1d1cd555086611c05df2b0b0c4b17b03065becb8d4fcd1bf9d2b66b32"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.779885 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c419-account-create-nxsz2" event={"ID":"80412713-2ef8-4499-ae54-2c6aebac5377","Type":"ContainerStarted","Data":"fa993fc897ccfb401916f1c31fde690941279a31a9c6fd061af3a720d635914f"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.784974 5035 generic.go:334] "Generic (PLEG): container finished" podID="8d84b9e7-688c-4926-bdce-1581f89d4875" containerID="5a6251900743edf35f1d5fcf6eba33637668b21e2d087572af49e2366a4d91a3" exitCode=0 Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.785039 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d325-account-create-dwc79" event={"ID":"8d84b9e7-688c-4926-bdce-1581f89d4875","Type":"ContainerDied","Data":"5a6251900743edf35f1d5fcf6eba33637668b21e2d087572af49e2366a4d91a3"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.785076 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d325-account-create-dwc79" event={"ID":"8d84b9e7-688c-4926-bdce-1581f89d4875","Type":"ContainerStarted","Data":"55c3890b93cb64bb8ee32d85e9e0fac7dc9990432d6f2dc2bb452c05b651e269"} Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.785230 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-cxbtl" podStartSLOduration=1.7852092819999998 podStartE2EDuration="1.785209282s" podCreationTimestamp="2025-11-24 19:32:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:32:11.78366503 +0000 UTC m=+1070.306171307" watchObservedRunningTime="2025-11-24 19:32:11.785209282 +0000 UTC m=+1070.307715539" Nov 24 19:32:11 crc kubenswrapper[5035]: I1124 19:32:11.822026 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f85b-account-create-zpj9j" podStartSLOduration=1.8220058689999998 podStartE2EDuration="1.822005869s" podCreationTimestamp="2025-11-24 19:32:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:32:11.817879826 +0000 UTC m=+1070.340386103" watchObservedRunningTime="2025-11-24 19:32:11.822005869 +0000 UTC m=+1070.344512126" Nov 24 19:32:12 crc kubenswrapper[5035]: I1124 19:32:12.801638 5035 generic.go:334] "Generic (PLEG): container finished" podID="da1e9bbe-02f8-4eda-be3d-bf882329f44c" containerID="2b65dff80626f7d183a900f4cf6dc685afc94965cc5638d12d4419aa23d9d680" exitCode=0 Nov 24 19:32:12 crc kubenswrapper[5035]: I1124 19:32:12.801736 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cxbtl" event={"ID":"da1e9bbe-02f8-4eda-be3d-bf882329f44c","Type":"ContainerDied","Data":"2b65dff80626f7d183a900f4cf6dc685afc94965cc5638d12d4419aa23d9d680"} Nov 24 19:32:12 crc kubenswrapper[5035]: I1124 19:32:12.806971 5035 generic.go:334] "Generic (PLEG): container finished" podID="aeff8377-e685-4e82-ad78-b9b99f9e1bc0" containerID="805534dd8880ebe48abf9b3a13b871ab38cd236f5d2fb19d4e1a9388254c77d9" exitCode=0 Nov 24 19:32:12 crc kubenswrapper[5035]: I1124 19:32:12.807123 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f85b-account-create-zpj9j" event={"ID":"aeff8377-e685-4e82-ad78-b9b99f9e1bc0","Type":"ContainerDied","Data":"805534dd8880ebe48abf9b3a13b871ab38cd236f5d2fb19d4e1a9388254c77d9"} Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.247322 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.261908 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.267812 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.279433 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383211 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d84b9e7-688c-4926-bdce-1581f89d4875-operator-scripts\") pod \"8d84b9e7-688c-4926-bdce-1581f89d4875\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383279 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx99t\" (UniqueName: \"kubernetes.io/projected/8d84b9e7-688c-4926-bdce-1581f89d4875-kube-api-access-zx99t\") pod \"8d84b9e7-688c-4926-bdce-1581f89d4875\" (UID: \"8d84b9e7-688c-4926-bdce-1581f89d4875\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383375 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btg56\" (UniqueName: \"kubernetes.io/projected/e080c702-5f10-4d8a-a192-07466d0d7644-kube-api-access-btg56\") pod \"e080c702-5f10-4d8a-a192-07466d0d7644\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383397 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb565\" (UniqueName: \"kubernetes.io/projected/80412713-2ef8-4499-ae54-2c6aebac5377-kube-api-access-zb565\") pod \"80412713-2ef8-4499-ae54-2c6aebac5377\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383417 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80412713-2ef8-4499-ae54-2c6aebac5377-operator-scripts\") pod \"80412713-2ef8-4499-ae54-2c6aebac5377\" (UID: \"80412713-2ef8-4499-ae54-2c6aebac5377\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383488 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68fcf962-da19-4161-8eee-aecde6615ba4-operator-scripts\") pod \"68fcf962-da19-4161-8eee-aecde6615ba4\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383560 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e080c702-5f10-4d8a-a192-07466d0d7644-operator-scripts\") pod \"e080c702-5f10-4d8a-a192-07466d0d7644\" (UID: \"e080c702-5f10-4d8a-a192-07466d0d7644\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383576 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x64hg\" (UniqueName: \"kubernetes.io/projected/68fcf962-da19-4161-8eee-aecde6615ba4-kube-api-access-x64hg\") pod \"68fcf962-da19-4161-8eee-aecde6615ba4\" (UID: \"68fcf962-da19-4161-8eee-aecde6615ba4\") " Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.383756 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d84b9e7-688c-4926-bdce-1581f89d4875-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d84b9e7-688c-4926-bdce-1581f89d4875" (UID: "8d84b9e7-688c-4926-bdce-1581f89d4875"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.384123 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80412713-2ef8-4499-ae54-2c6aebac5377-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80412713-2ef8-4499-ae54-2c6aebac5377" (UID: "80412713-2ef8-4499-ae54-2c6aebac5377"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.384225 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68fcf962-da19-4161-8eee-aecde6615ba4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68fcf962-da19-4161-8eee-aecde6615ba4" (UID: "68fcf962-da19-4161-8eee-aecde6615ba4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.384377 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e080c702-5f10-4d8a-a192-07466d0d7644-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e080c702-5f10-4d8a-a192-07466d0d7644" (UID: "e080c702-5f10-4d8a-a192-07466d0d7644"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.384690 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80412713-2ef8-4499-ae54-2c6aebac5377-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.384703 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68fcf962-da19-4161-8eee-aecde6615ba4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.384714 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e080c702-5f10-4d8a-a192-07466d0d7644-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.384725 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d84b9e7-688c-4926-bdce-1581f89d4875-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.396489 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d84b9e7-688c-4926-bdce-1581f89d4875-kube-api-access-zx99t" (OuterVolumeSpecName: "kube-api-access-zx99t") pod "8d84b9e7-688c-4926-bdce-1581f89d4875" (UID: "8d84b9e7-688c-4926-bdce-1581f89d4875"). InnerVolumeSpecName "kube-api-access-zx99t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.396570 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e080c702-5f10-4d8a-a192-07466d0d7644-kube-api-access-btg56" (OuterVolumeSpecName: "kube-api-access-btg56") pod "e080c702-5f10-4d8a-a192-07466d0d7644" (UID: "e080c702-5f10-4d8a-a192-07466d0d7644"). InnerVolumeSpecName "kube-api-access-btg56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.396688 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fcf962-da19-4161-8eee-aecde6615ba4-kube-api-access-x64hg" (OuterVolumeSpecName: "kube-api-access-x64hg") pod "68fcf962-da19-4161-8eee-aecde6615ba4" (UID: "68fcf962-da19-4161-8eee-aecde6615ba4"). InnerVolumeSpecName "kube-api-access-x64hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.396752 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80412713-2ef8-4499-ae54-2c6aebac5377-kube-api-access-zb565" (OuterVolumeSpecName: "kube-api-access-zb565") pod "80412713-2ef8-4499-ae54-2c6aebac5377" (UID: "80412713-2ef8-4499-ae54-2c6aebac5377"). InnerVolumeSpecName "kube-api-access-zb565". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.485929 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx99t\" (UniqueName: \"kubernetes.io/projected/8d84b9e7-688c-4926-bdce-1581f89d4875-kube-api-access-zx99t\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.485967 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btg56\" (UniqueName: \"kubernetes.io/projected/e080c702-5f10-4d8a-a192-07466d0d7644-kube-api-access-btg56\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.485977 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb565\" (UniqueName: \"kubernetes.io/projected/80412713-2ef8-4499-ae54-2c6aebac5377-kube-api-access-zb565\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.485986 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x64hg\" (UniqueName: \"kubernetes.io/projected/68fcf962-da19-4161-8eee-aecde6615ba4-kube-api-access-x64hg\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.832011 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c419-account-create-nxsz2" event={"ID":"80412713-2ef8-4499-ae54-2c6aebac5377","Type":"ContainerDied","Data":"fa993fc897ccfb401916f1c31fde690941279a31a9c6fd061af3a720d635914f"} Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.832055 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa993fc897ccfb401916f1c31fde690941279a31a9c6fd061af3a720d635914f" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.832133 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c419-account-create-nxsz2" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.841659 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d325-account-create-dwc79" event={"ID":"8d84b9e7-688c-4926-bdce-1581f89d4875","Type":"ContainerDied","Data":"55c3890b93cb64bb8ee32d85e9e0fac7dc9990432d6f2dc2bb452c05b651e269"} Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.841704 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55c3890b93cb64bb8ee32d85e9e0fac7dc9990432d6f2dc2bb452c05b651e269" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.841787 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d325-account-create-dwc79" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.855155 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vrrsl" event={"ID":"e080c702-5f10-4d8a-a192-07466d0d7644","Type":"ContainerDied","Data":"4e9f9be866f32e784880811da2b4a009fcc952108c9760ed8ee31c4f4f7942c0"} Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.855613 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e9f9be866f32e784880811da2b4a009fcc952108c9760ed8ee31c4f4f7942c0" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.855217 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vrrsl" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.874019 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dgz8h" Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.874887 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dgz8h" event={"ID":"68fcf962-da19-4161-8eee-aecde6615ba4","Type":"ContainerDied","Data":"07b0f55bef9eef64887643db4f6540ace9090e080678b1f78c338a0bd5c69006"} Nov 24 19:32:13 crc kubenswrapper[5035]: I1124 19:32:13.875017 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07b0f55bef9eef64887643db4f6540ace9090e080678b1f78c338a0bd5c69006" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.748310 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.751560 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.907907 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-operator-scripts\") pod \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.908007 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da1e9bbe-02f8-4eda-be3d-bf882329f44c-operator-scripts\") pod \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.908139 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4gmr\" (UniqueName: \"kubernetes.io/projected/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-kube-api-access-j4gmr\") pod \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\" (UID: \"aeff8377-e685-4e82-ad78-b9b99f9e1bc0\") " Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.908233 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqwnp\" (UniqueName: \"kubernetes.io/projected/da1e9bbe-02f8-4eda-be3d-bf882329f44c-kube-api-access-qqwnp\") pod \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\" (UID: \"da1e9bbe-02f8-4eda-be3d-bf882329f44c\") " Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.909122 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aeff8377-e685-4e82-ad78-b9b99f9e1bc0" (UID: "aeff8377-e685-4e82-ad78-b9b99f9e1bc0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.909897 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da1e9bbe-02f8-4eda-be3d-bf882329f44c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "da1e9bbe-02f8-4eda-be3d-bf882329f44c" (UID: "da1e9bbe-02f8-4eda-be3d-bf882329f44c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.916870 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da1e9bbe-02f8-4eda-be3d-bf882329f44c-kube-api-access-qqwnp" (OuterVolumeSpecName: "kube-api-access-qqwnp") pod "da1e9bbe-02f8-4eda-be3d-bf882329f44c" (UID: "da1e9bbe-02f8-4eda-be3d-bf882329f44c"). InnerVolumeSpecName "kube-api-access-qqwnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.917470 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-kube-api-access-j4gmr" (OuterVolumeSpecName: "kube-api-access-j4gmr") pod "aeff8377-e685-4e82-ad78-b9b99f9e1bc0" (UID: "aeff8377-e685-4e82-ad78-b9b99f9e1bc0"). InnerVolumeSpecName "kube-api-access-j4gmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.927225 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cxbtl" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.927239 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cxbtl" event={"ID":"da1e9bbe-02f8-4eda-be3d-bf882329f44c","Type":"ContainerDied","Data":"7e03111bb7eb58f5066cd291710d626ce3ef6206fca58440033368c0b96275a1"} Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.927338 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e03111bb7eb58f5066cd291710d626ce3ef6206fca58440033368c0b96275a1" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.930353 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f85b-account-create-zpj9j" event={"ID":"aeff8377-e685-4e82-ad78-b9b99f9e1bc0","Type":"ContainerDied","Data":"9952a67c048da2ff42b23962a89a63e15d4c12c1112e67ad64a7b8eb98f3c7aa"} Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.930401 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9952a67c048da2ff42b23962a89a63e15d4c12c1112e67ad64a7b8eb98f3c7aa" Nov 24 19:32:19 crc kubenswrapper[5035]: I1124 19:32:19.930476 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f85b-account-create-zpj9j" Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.010579 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4gmr\" (UniqueName: \"kubernetes.io/projected/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-kube-api-access-j4gmr\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.010862 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqwnp\" (UniqueName: \"kubernetes.io/projected/da1e9bbe-02f8-4eda-be3d-bf882329f44c-kube-api-access-qqwnp\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.010873 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aeff8377-e685-4e82-ad78-b9b99f9e1bc0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.010882 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da1e9bbe-02f8-4eda-be3d-bf882329f44c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.940786 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9jwdn" event={"ID":"7ba53ebf-bae3-4629-af0b-8637e6491515","Type":"ContainerStarted","Data":"21dabb6c2c2e93d18439131b14ef79c29831cb76666b75d28f8ff63a3ce15098"} Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.942725 5035 generic.go:334] "Generic (PLEG): container finished" podID="030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" containerID="4cce725c77ece9593b000a253ae81b3eb7df00c2a24ec012d0b1f9d34272c134" exitCode=0 Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.942769 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qrcbh" event={"ID":"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe","Type":"ContainerDied","Data":"4cce725c77ece9593b000a253ae81b3eb7df00c2a24ec012d0b1f9d34272c134"} Nov 24 19:32:20 crc kubenswrapper[5035]: I1124 19:32:20.964085 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-9jwdn" podStartSLOduration=2.212505294 podStartE2EDuration="10.964063651s" podCreationTimestamp="2025-11-24 19:32:10 +0000 UTC" firstStartedPulling="2025-11-24 19:32:11.154883751 +0000 UTC m=+1069.677390008" lastFinishedPulling="2025-11-24 19:32:19.906442078 +0000 UTC m=+1078.428948365" observedRunningTime="2025-11-24 19:32:20.958862588 +0000 UTC m=+1079.481368855" watchObservedRunningTime="2025-11-24 19:32:20.964063651 +0000 UTC m=+1079.486569918" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.376720 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qrcbh" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.454249 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-config-data\") pod \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.454646 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-combined-ca-bundle\") pod \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.454682 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-db-sync-config-data\") pod \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.454918 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56kqh\" (UniqueName: \"kubernetes.io/projected/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-kube-api-access-56kqh\") pod \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\" (UID: \"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe\") " Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.460042 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" (UID: "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.460914 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-kube-api-access-56kqh" (OuterVolumeSpecName: "kube-api-access-56kqh") pod "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" (UID: "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe"). InnerVolumeSpecName "kube-api-access-56kqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.490623 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" (UID: "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.492893 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-config-data" (OuterVolumeSpecName: "config-data") pod "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" (UID: "030eac16-e8f0-4a3c-9d1e-588cc25d6ebe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.557146 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.557183 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.557197 5035 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.557206 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56kqh\" (UniqueName: \"kubernetes.io/projected/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe-kube-api-access-56kqh\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.984373 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qrcbh" Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.984417 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qrcbh" event={"ID":"030eac16-e8f0-4a3c-9d1e-588cc25d6ebe","Type":"ContainerDied","Data":"ebdfe74cc37e9c81bea6d7d08b3b1d5efe8eceefe93cc2ab0299cfc91f59b9d9"} Nov 24 19:32:22 crc kubenswrapper[5035]: I1124 19:32:22.984484 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebdfe74cc37e9c81bea6d7d08b3b1d5efe8eceefe93cc2ab0299cfc91f59b9d9" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.394617 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75b58765b5-pvzps"] Nov 24 19:32:23 crc kubenswrapper[5035]: E1124 19:32:23.394993 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80412713-2ef8-4499-ae54-2c6aebac5377" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395009 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="80412713-2ef8-4499-ae54-2c6aebac5377" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: E1124 19:32:23.395018 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fcf962-da19-4161-8eee-aecde6615ba4" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395025 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fcf962-da19-4161-8eee-aecde6615ba4" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: E1124 19:32:23.395039 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e080c702-5f10-4d8a-a192-07466d0d7644" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395046 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e080c702-5f10-4d8a-a192-07466d0d7644" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: E1124 19:32:23.395060 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" containerName="glance-db-sync" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395066 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" containerName="glance-db-sync" Nov 24 19:32:23 crc kubenswrapper[5035]: E1124 19:32:23.395076 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d84b9e7-688c-4926-bdce-1581f89d4875" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395081 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d84b9e7-688c-4926-bdce-1581f89d4875" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: E1124 19:32:23.395092 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1e9bbe-02f8-4eda-be3d-bf882329f44c" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395098 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1e9bbe-02f8-4eda-be3d-bf882329f44c" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: E1124 19:32:23.395112 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeff8377-e685-4e82-ad78-b9b99f9e1bc0" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395118 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeff8377-e685-4e82-ad78-b9b99f9e1bc0" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395308 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="da1e9bbe-02f8-4eda-be3d-bf882329f44c" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395321 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeff8377-e685-4e82-ad78-b9b99f9e1bc0" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395336 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e080c702-5f10-4d8a-a192-07466d0d7644" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395344 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" containerName="glance-db-sync" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395354 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fcf962-da19-4161-8eee-aecde6615ba4" containerName="mariadb-database-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395362 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d84b9e7-688c-4926-bdce-1581f89d4875" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.395368 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="80412713-2ef8-4499-ae54-2c6aebac5377" containerName="mariadb-account-create" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.396250 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.414226 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75b58765b5-pvzps"] Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.472572 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-config\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.472650 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-dns-svc\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.472745 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-nb\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.472798 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-sb\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.472828 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xckx6\" (UniqueName: \"kubernetes.io/projected/9599d9a7-9271-4c06-944f-4ce340e70ab6-kube-api-access-xckx6\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.573923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-dns-svc\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.574011 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-nb\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.574047 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-sb\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.574069 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xckx6\" (UniqueName: \"kubernetes.io/projected/9599d9a7-9271-4c06-944f-4ce340e70ab6-kube-api-access-xckx6\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.574108 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-config\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.574964 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-dns-svc\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.575034 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-config\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.575034 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-nb\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.575113 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-sb\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.592237 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xckx6\" (UniqueName: \"kubernetes.io/projected/9599d9a7-9271-4c06-944f-4ce340e70ab6-kube-api-access-xckx6\") pod \"dnsmasq-dns-75b58765b5-pvzps\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.713180 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.934213 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75b58765b5-pvzps"] Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.993337 5035 generic.go:334] "Generic (PLEG): container finished" podID="7ba53ebf-bae3-4629-af0b-8637e6491515" containerID="21dabb6c2c2e93d18439131b14ef79c29831cb76666b75d28f8ff63a3ce15098" exitCode=0 Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.993412 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9jwdn" event={"ID":"7ba53ebf-bae3-4629-af0b-8637e6491515","Type":"ContainerDied","Data":"21dabb6c2c2e93d18439131b14ef79c29831cb76666b75d28f8ff63a3ce15098"} Nov 24 19:32:23 crc kubenswrapper[5035]: I1124 19:32:23.994775 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" event={"ID":"9599d9a7-9271-4c06-944f-4ce340e70ab6","Type":"ContainerStarted","Data":"c237af6029c5f2c6b319c77600a479c1e43f1f487185a16ad3f800394ec64d92"} Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.008337 5035 generic.go:334] "Generic (PLEG): container finished" podID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerID="f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86" exitCode=0 Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.008566 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" event={"ID":"9599d9a7-9271-4c06-944f-4ce340e70ab6","Type":"ContainerDied","Data":"f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86"} Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.293471 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.403126 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qdvg\" (UniqueName: \"kubernetes.io/projected/7ba53ebf-bae3-4629-af0b-8637e6491515-kube-api-access-2qdvg\") pod \"7ba53ebf-bae3-4629-af0b-8637e6491515\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.403206 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-config-data\") pod \"7ba53ebf-bae3-4629-af0b-8637e6491515\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.403274 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-combined-ca-bundle\") pod \"7ba53ebf-bae3-4629-af0b-8637e6491515\" (UID: \"7ba53ebf-bae3-4629-af0b-8637e6491515\") " Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.407587 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ba53ebf-bae3-4629-af0b-8637e6491515-kube-api-access-2qdvg" (OuterVolumeSpecName: "kube-api-access-2qdvg") pod "7ba53ebf-bae3-4629-af0b-8637e6491515" (UID: "7ba53ebf-bae3-4629-af0b-8637e6491515"). InnerVolumeSpecName "kube-api-access-2qdvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.425812 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ba53ebf-bae3-4629-af0b-8637e6491515" (UID: "7ba53ebf-bae3-4629-af0b-8637e6491515"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.440909 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-config-data" (OuterVolumeSpecName: "config-data") pod "7ba53ebf-bae3-4629-af0b-8637e6491515" (UID: "7ba53ebf-bae3-4629-af0b-8637e6491515"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.505099 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.505141 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qdvg\" (UniqueName: \"kubernetes.io/projected/7ba53ebf-bae3-4629-af0b-8637e6491515-kube-api-access-2qdvg\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:25 crc kubenswrapper[5035]: I1124 19:32:25.505157 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba53ebf-bae3-4629-af0b-8637e6491515-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.022435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9jwdn" event={"ID":"7ba53ebf-bae3-4629-af0b-8637e6491515","Type":"ContainerDied","Data":"70bcb384ad3f7471aac600c64360ffd39116654b4ec3fda042e1e532de1c0d5c"} Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.022791 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70bcb384ad3f7471aac600c64360ffd39116654b4ec3fda042e1e532de1c0d5c" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.022869 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9jwdn" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.030600 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" event={"ID":"9599d9a7-9271-4c06-944f-4ce340e70ab6","Type":"ContainerStarted","Data":"272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a"} Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.030809 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.066714 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" podStartSLOduration=3.066695324 podStartE2EDuration="3.066695324s" podCreationTimestamp="2025-11-24 19:32:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:32:26.055453326 +0000 UTC m=+1084.577959623" watchObservedRunningTime="2025-11-24 19:32:26.066695324 +0000 UTC m=+1084.589201581" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.230543 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75b58765b5-pvzps"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.285466 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d44dbddd5-jgjfz"] Nov 24 19:32:26 crc kubenswrapper[5035]: E1124 19:32:26.285814 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba53ebf-bae3-4629-af0b-8637e6491515" containerName="keystone-db-sync" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.285825 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba53ebf-bae3-4629-af0b-8637e6491515" containerName="keystone-db-sync" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.285976 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ba53ebf-bae3-4629-af0b-8637e6491515" containerName="keystone-db-sync" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.286802 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.294503 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sm9mf"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.295659 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.297500 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.297751 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.298680 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.298996 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.299143 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2trk2" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.357689 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d44dbddd5-jgjfz"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.374015 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sm9mf"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.421853 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-combined-ca-bundle\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.421912 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-dns-svc\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.421939 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-sb\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.421978 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-fernet-keys\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.422021 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-config-data\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.422049 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-credential-keys\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.422086 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78mrs\" (UniqueName: \"kubernetes.io/projected/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-kube-api-access-78mrs\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.422110 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-scripts\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.422135 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-config\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.422177 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfmn6\" (UniqueName: \"kubernetes.io/projected/961faa76-0c57-4ec1-891e-520dc13230bd-kube-api-access-bfmn6\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.422221 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-nb\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.449104 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c8bf8bb97-lzjsb"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.450790 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.457967 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-4xhrl" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.458159 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.458277 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.458550 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.523873 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-fernet-keys\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.523939 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f6391e26-277d-4d6e-bbf1-82046cbfe80f-horizon-secret-key\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.523967 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfzv5\" (UniqueName: \"kubernetes.io/projected/f6391e26-277d-4d6e-bbf1-82046cbfe80f-kube-api-access-zfzv5\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.523993 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-config-data\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524017 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6391e26-277d-4d6e-bbf1-82046cbfe80f-logs\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524053 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-credential-keys\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524093 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78mrs\" (UniqueName: \"kubernetes.io/projected/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-kube-api-access-78mrs\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524110 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-scripts\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524130 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-config-data\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524148 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-config\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524181 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfmn6\" (UniqueName: \"kubernetes.io/projected/961faa76-0c57-4ec1-891e-520dc13230bd-kube-api-access-bfmn6\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524217 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-nb\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524242 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-scripts\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524263 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-combined-ca-bundle\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524280 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-dns-svc\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.524308 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-sb\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.525270 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-nb\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.525927 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-config\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.525987 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-sb\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.526696 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-dns-svc\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.544010 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-credential-keys\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.544058 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-fernet-keys\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.545473 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-scripts\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.553171 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-config-data\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.564568 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-combined-ca-bundle\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.568354 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c8bf8bb97-lzjsb"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.586401 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-g7nvj"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.587423 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.588456 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78mrs\" (UniqueName: \"kubernetes.io/projected/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-kube-api-access-78mrs\") pod \"keystone-bootstrap-sm9mf\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.591343 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfmn6\" (UniqueName: \"kubernetes.io/projected/961faa76-0c57-4ec1-891e-520dc13230bd-kube-api-access-bfmn6\") pod \"dnsmasq-dns-5d44dbddd5-jgjfz\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.596218 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.596543 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qbgkv" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.596787 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.616842 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.645732 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646331 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-scripts\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646391 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f6391e26-277d-4d6e-bbf1-82046cbfe80f-horizon-secret-key\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646415 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfzv5\" (UniqueName: \"kubernetes.io/projected/f6391e26-277d-4d6e-bbf1-82046cbfe80f-kube-api-access-zfzv5\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646442 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6391e26-277d-4d6e-bbf1-82046cbfe80f-logs\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646461 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-config\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646503 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-config-data\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646525 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqksl\" (UniqueName: \"kubernetes.io/projected/c11a04cb-bd10-49c8-affa-5189af979040-kube-api-access-tqksl\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.646566 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-combined-ca-bundle\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.647517 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-scripts\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.649206 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6391e26-277d-4d6e-bbf1-82046cbfe80f-logs\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.655179 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-config-data\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.656325 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f6391e26-277d-4d6e-bbf1-82046cbfe80f-horizon-secret-key\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.665005 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-nbbjc"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.666034 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.685807 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5mm55" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.686101 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.686233 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.686335 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g7nvj"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.700547 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfzv5\" (UniqueName: \"kubernetes.io/projected/f6391e26-277d-4d6e-bbf1-82046cbfe80f-kube-api-access-zfzv5\") pod \"horizon-7c8bf8bb97-lzjsb\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.726454 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nbbjc"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750124 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-etc-machine-id\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750191 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-config\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750218 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-config-data\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750243 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-scripts\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750322 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqksl\" (UniqueName: \"kubernetes.io/projected/c11a04cb-bd10-49c8-affa-5189af979040-kube-api-access-tqksl\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750369 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-combined-ca-bundle\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750401 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-combined-ca-bundle\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750425 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-db-sync-config-data\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.750455 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl5n6\" (UniqueName: \"kubernetes.io/projected/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-kube-api-access-dl5n6\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.774732 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.776658 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-config\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.779484 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.781533 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.791843 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-combined-ca-bundle\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.792404 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.792657 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.811141 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqksl\" (UniqueName: \"kubernetes.io/projected/c11a04cb-bd10-49c8-affa-5189af979040-kube-api-access-tqksl\") pod \"neutron-db-sync-g7nvj\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.825418 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.850723 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-h76mn"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.851891 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.851951 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-combined-ca-bundle\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.851979 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-db-sync-config-data\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852003 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl5n6\" (UniqueName: \"kubernetes.io/projected/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-kube-api-access-dl5n6\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852028 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-log-httpd\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852062 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-run-httpd\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852086 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852113 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-etc-machine-id\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852137 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-config-data\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852160 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-scripts\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852238 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-config-data\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852314 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-scripts\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852339 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7md9\" (UniqueName: \"kubernetes.io/projected/f9739bc1-965c-4f67-9b14-b96e176d72c1-kube-api-access-k7md9\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.852989 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-etc-machine-id\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.862722 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-scripts\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.862849 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-config-data\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.864179 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.875784 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-db-sync-config-data\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.877730 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.877864 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-combined-ca-bundle\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.878082 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sqtx8" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.878307 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.899398 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl5n6\" (UniqueName: \"kubernetes.io/projected/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-kube-api-access-dl5n6\") pod \"cinder-db-sync-nbbjc\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.917396 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xb7gt"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.918819 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.931862 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-v6rv7" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.932077 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.946876 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.950958 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-h76mn"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953613 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-log-httpd\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953670 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953689 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-run-httpd\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953712 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-combined-ca-bundle\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953731 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-scripts\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953756 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvlqp\" (UniqueName: \"kubernetes.io/projected/764ce995-4314-4c29-8b00-dc264344edf0-kube-api-access-mvlqp\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953803 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-config-data\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953823 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-scripts\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953841 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7md9\" (UniqueName: \"kubernetes.io/projected/f9739bc1-965c-4f67-9b14-b96e176d72c1-kube-api-access-k7md9\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953860 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953876 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/764ce995-4314-4c29-8b00-dc264344edf0-logs\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.953897 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-config-data\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.954401 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-log-httpd\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.954735 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-run-httpd\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.964526 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d44dbddd5-jgjfz"] Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.973056 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-scripts\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.973474 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-config-data\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.978047 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.978723 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.981157 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:26 crc kubenswrapper[5035]: I1124 19:32:26.999849 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xb7gt"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.001150 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7md9\" (UniqueName: \"kubernetes.io/projected/f9739bc1-965c-4f67-9b14-b96e176d72c1-kube-api-access-k7md9\") pod \"ceilometer-0\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " pod="openstack/ceilometer-0" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.011820 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.028699 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f8f5cc67-rtdk2"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.029999 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.052350 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f8f5cc67-rtdk2"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055313 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-combined-ca-bundle\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055377 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/764ce995-4314-4c29-8b00-dc264344edf0-logs\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055407 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-config-data\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055428 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ckjt\" (UniqueName: \"kubernetes.io/projected/205bbe21-12fb-4cdf-bf2d-09008085f221-kube-api-access-6ckjt\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055488 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-db-sync-config-data\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055555 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-combined-ca-bundle\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055573 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-scripts\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.055594 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvlqp\" (UniqueName: \"kubernetes.io/projected/764ce995-4314-4c29-8b00-dc264344edf0-kube-api-access-mvlqp\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.056477 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/764ce995-4314-4c29-8b00-dc264344edf0-logs\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.062715 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d46ff7b7c-jf8nx"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.065514 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.069041 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d46ff7b7c-jf8nx"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.071832 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-scripts\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.075733 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-combined-ca-bundle\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.099759 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvlqp\" (UniqueName: \"kubernetes.io/projected/764ce995-4314-4c29-8b00-dc264344edf0-kube-api-access-mvlqp\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.104924 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-config-data\") pod \"placement-db-sync-h76mn\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158063 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae039f5-86cf-46b9-a9c4-c3a18849bfec-logs\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158143 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-db-sync-config-data\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158181 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mgk2\" (UniqueName: \"kubernetes.io/projected/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-kube-api-access-6mgk2\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158257 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-config-data\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158328 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-nb\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158351 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-sb\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158405 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-scripts\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158503 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aae039f5-86cf-46b9-a9c4-c3a18849bfec-horizon-secret-key\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158566 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7www\" (UniqueName: \"kubernetes.io/projected/aae039f5-86cf-46b9-a9c4-c3a18849bfec-kube-api-access-s7www\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158608 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-config\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158634 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-combined-ca-bundle\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158690 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ckjt\" (UniqueName: \"kubernetes.io/projected/205bbe21-12fb-4cdf-bf2d-09008085f221-kube-api-access-6ckjt\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.158778 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-dns-svc\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.162779 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-db-sync-config-data\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.166964 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-combined-ca-bundle\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.182680 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ckjt\" (UniqueName: \"kubernetes.io/projected/205bbe21-12fb-4cdf-bf2d-09008085f221-kube-api-access-6ckjt\") pod \"barbican-db-sync-xb7gt\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259680 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-dns-svc\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259720 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae039f5-86cf-46b9-a9c4-c3a18849bfec-logs\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mgk2\" (UniqueName: \"kubernetes.io/projected/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-kube-api-access-6mgk2\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259780 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-config-data\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259810 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-nb\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259825 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-sb\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259845 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-scripts\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259907 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aae039f5-86cf-46b9-a9c4-c3a18849bfec-horizon-secret-key\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259925 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7www\" (UniqueName: \"kubernetes.io/projected/aae039f5-86cf-46b9-a9c4-c3a18849bfec-kube-api-access-s7www\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.259948 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-config\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.260663 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-config\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.261158 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-dns-svc\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.261408 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae039f5-86cf-46b9-a9c4-c3a18849bfec-logs\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.262278 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-scripts\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.263115 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-nb\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.263237 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-sb\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.263479 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-config-data\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.273141 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aae039f5-86cf-46b9-a9c4-c3a18849bfec-horizon-secret-key\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.279996 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mgk2\" (UniqueName: \"kubernetes.io/projected/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-kube-api-access-6mgk2\") pod \"dnsmasq-dns-7f8f5cc67-rtdk2\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.282905 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7www\" (UniqueName: \"kubernetes.io/projected/aae039f5-86cf-46b9-a9c4-c3a18849bfec-kube-api-access-s7www\") pod \"horizon-d46ff7b7c-jf8nx\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.344498 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h76mn" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.375847 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.401174 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.426593 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.601978 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sm9mf"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.611346 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d44dbddd5-jgjfz"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.615417 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c8bf8bb97-lzjsb"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.773774 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nbbjc"] Nov 24 19:32:27 crc kubenswrapper[5035]: W1124 19:32:27.778207 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c9794ab_e4e3_4af0_a689_6ef761d1fe4f.slice/crio-fd69585e1eb719c35886ba4568491c7b02dc4dc0456e059e4165bf77dde01e19 WatchSource:0}: Error finding container fd69585e1eb719c35886ba4568491c7b02dc4dc0456e059e4165bf77dde01e19: Status 404 returned error can't find the container with id fd69585e1eb719c35886ba4568491c7b02dc4dc0456e059e4165bf77dde01e19 Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.782799 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-g7nvj"] Nov 24 19:32:27 crc kubenswrapper[5035]: I1124 19:32:27.906250 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:32:28 crc kubenswrapper[5035]: W1124 19:32:28.036710 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaae039f5_86cf_46b9_a9c4_c3a18849bfec.slice/crio-473944120fc3108dfc9deb80eda5e95ce087929ac2d5a8df9186602df8bcb910 WatchSource:0}: Error finding container 473944120fc3108dfc9deb80eda5e95ce087929ac2d5a8df9186602df8bcb910: Status 404 returned error can't find the container with id 473944120fc3108dfc9deb80eda5e95ce087929ac2d5a8df9186602df8bcb910 Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.051906 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d46ff7b7c-jf8nx"] Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.065525 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" event={"ID":"961faa76-0c57-4ec1-891e-520dc13230bd","Type":"ContainerStarted","Data":"b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.065579 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" event={"ID":"961faa76-0c57-4ec1-891e-520dc13230bd","Type":"ContainerStarted","Data":"0c22f86dd9b38ceaeefc917998c04bc4b01351565334ef582a1815b910b770d5"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.067963 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8bf8bb97-lzjsb" event={"ID":"f6391e26-277d-4d6e-bbf1-82046cbfe80f","Type":"ContainerStarted","Data":"3616bf401bb3c4a536edf5411bc9ab95375cd152ef62267aa73f3e0a99fe0125"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.086253 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g7nvj" event={"ID":"c11a04cb-bd10-49c8-affa-5189af979040","Type":"ContainerStarted","Data":"4c6bb91f085a9a7f1b6c300c57b5ba3691a074db9a73773a414462dcd34bb968"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.086317 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g7nvj" event={"ID":"c11a04cb-bd10-49c8-affa-5189af979040","Type":"ContainerStarted","Data":"cae340075a47e28d8ed89b9c6b4a4c708894e1f900d3f92d54a523155e84e9c9"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.091147 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbbjc" event={"ID":"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f","Type":"ContainerStarted","Data":"fd69585e1eb719c35886ba4568491c7b02dc4dc0456e059e4165bf77dde01e19"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.094241 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d46ff7b7c-jf8nx" event={"ID":"aae039f5-86cf-46b9-a9c4-c3a18849bfec","Type":"ContainerStarted","Data":"473944120fc3108dfc9deb80eda5e95ce087929ac2d5a8df9186602df8bcb910"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.098376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sm9mf" event={"ID":"5bfe89ec-3b9c-4999-9955-d3769ce1caa2","Type":"ContainerStarted","Data":"f9d14da07ccc04044087429f74e3a615ac05903701b026bdc52ac7caa9d187e3"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.098418 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sm9mf" event={"ID":"5bfe89ec-3b9c-4999-9955-d3769ce1caa2","Type":"ContainerStarted","Data":"c6500b660008ffbb72cf19abd7c43ae9f7cfb81bf5eec1e09e48607ec63f9f94"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.100137 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" podUID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerName="dnsmasq-dns" containerID="cri-o://272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a" gracePeriod=10 Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.100404 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerStarted","Data":"fb54ffc05444118e954d391982ce8bc1d1485cd98d3b5f37dd69b0255068a973"} Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.109217 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-h76mn"] Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.127711 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-g7nvj" podStartSLOduration=2.127691743 podStartE2EDuration="2.127691743s" podCreationTimestamp="2025-11-24 19:32:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:32:28.108952481 +0000 UTC m=+1086.631458738" watchObservedRunningTime="2025-11-24 19:32:28.127691743 +0000 UTC m=+1086.650198000" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.131986 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sm9mf" podStartSLOduration=2.13196587 podStartE2EDuration="2.13196587s" podCreationTimestamp="2025-11-24 19:32:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:32:28.122715527 +0000 UTC m=+1086.645221784" watchObservedRunningTime="2025-11-24 19:32:28.13196587 +0000 UTC m=+1086.654472127" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.172326 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f8f5cc67-rtdk2"] Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.178098 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xb7gt"] Nov 24 19:32:28 crc kubenswrapper[5035]: W1124 19:32:28.196209 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod205bbe21_12fb_4cdf_bf2d_09008085f221.slice/crio-1eb02672bfc9ed20132b6b3b0be8dfc5c2e03ed66cde46f6ae4c784372d6d179 WatchSource:0}: Error finding container 1eb02672bfc9ed20132b6b3b0be8dfc5c2e03ed66cde46f6ae4c784372d6d179: Status 404 returned error can't find the container with id 1eb02672bfc9ed20132b6b3b0be8dfc5c2e03ed66cde46f6ae4c784372d6d179 Nov 24 19:32:28 crc kubenswrapper[5035]: W1124 19:32:28.197269 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b2bb2c4_3cf8_4385_a5a3_e8576b6a0e48.slice/crio-8b331ffbbc33898aabf40c6b2d1443db8ece40e8d576a189fa2f7f4c46bb3fd3 WatchSource:0}: Error finding container 8b331ffbbc33898aabf40c6b2d1443db8ece40e8d576a189fa2f7f4c46bb3fd3: Status 404 returned error can't find the container with id 8b331ffbbc33898aabf40c6b2d1443db8ece40e8d576a189fa2f7f4c46bb3fd3 Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.346131 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.491992 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfmn6\" (UniqueName: \"kubernetes.io/projected/961faa76-0c57-4ec1-891e-520dc13230bd-kube-api-access-bfmn6\") pod \"961faa76-0c57-4ec1-891e-520dc13230bd\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.492344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-dns-svc\") pod \"961faa76-0c57-4ec1-891e-520dc13230bd\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.492385 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-sb\") pod \"961faa76-0c57-4ec1-891e-520dc13230bd\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.492460 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-nb\") pod \"961faa76-0c57-4ec1-891e-520dc13230bd\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.492491 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-config\") pod \"961faa76-0c57-4ec1-891e-520dc13230bd\" (UID: \"961faa76-0c57-4ec1-891e-520dc13230bd\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.496983 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/961faa76-0c57-4ec1-891e-520dc13230bd-kube-api-access-bfmn6" (OuterVolumeSpecName: "kube-api-access-bfmn6") pod "961faa76-0c57-4ec1-891e-520dc13230bd" (UID: "961faa76-0c57-4ec1-891e-520dc13230bd"). InnerVolumeSpecName "kube-api-access-bfmn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.517194 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.518999 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-config" (OuterVolumeSpecName: "config") pod "961faa76-0c57-4ec1-891e-520dc13230bd" (UID: "961faa76-0c57-4ec1-891e-520dc13230bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.528876 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "961faa76-0c57-4ec1-891e-520dc13230bd" (UID: "961faa76-0c57-4ec1-891e-520dc13230bd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.531387 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "961faa76-0c57-4ec1-891e-520dc13230bd" (UID: "961faa76-0c57-4ec1-891e-520dc13230bd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.573280 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "961faa76-0c57-4ec1-891e-520dc13230bd" (UID: "961faa76-0c57-4ec1-891e-520dc13230bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.593783 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-sb\") pod \"9599d9a7-9271-4c06-944f-4ce340e70ab6\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.593843 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-dns-svc\") pod \"9599d9a7-9271-4c06-944f-4ce340e70ab6\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.593860 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-nb\") pod \"9599d9a7-9271-4c06-944f-4ce340e70ab6\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.593876 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-config\") pod \"9599d9a7-9271-4c06-944f-4ce340e70ab6\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.593946 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xckx6\" (UniqueName: \"kubernetes.io/projected/9599d9a7-9271-4c06-944f-4ce340e70ab6-kube-api-access-xckx6\") pod \"9599d9a7-9271-4c06-944f-4ce340e70ab6\" (UID: \"9599d9a7-9271-4c06-944f-4ce340e70ab6\") " Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.594284 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.594305 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.594315 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfmn6\" (UniqueName: \"kubernetes.io/projected/961faa76-0c57-4ec1-891e-520dc13230bd-kube-api-access-bfmn6\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.594323 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.594331 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/961faa76-0c57-4ec1-891e-520dc13230bd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.609276 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9599d9a7-9271-4c06-944f-4ce340e70ab6-kube-api-access-xckx6" (OuterVolumeSpecName: "kube-api-access-xckx6") pod "9599d9a7-9271-4c06-944f-4ce340e70ab6" (UID: "9599d9a7-9271-4c06-944f-4ce340e70ab6"). InnerVolumeSpecName "kube-api-access-xckx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.658035 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9599d9a7-9271-4c06-944f-4ce340e70ab6" (UID: "9599d9a7-9271-4c06-944f-4ce340e70ab6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.662660 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9599d9a7-9271-4c06-944f-4ce340e70ab6" (UID: "9599d9a7-9271-4c06-944f-4ce340e70ab6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.668950 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9599d9a7-9271-4c06-944f-4ce340e70ab6" (UID: "9599d9a7-9271-4c06-944f-4ce340e70ab6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.677738 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-config" (OuterVolumeSpecName: "config") pod "9599d9a7-9271-4c06-944f-4ce340e70ab6" (UID: "9599d9a7-9271-4c06-944f-4ce340e70ab6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.699604 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.699776 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.699822 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.699838 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xckx6\" (UniqueName: \"kubernetes.io/projected/9599d9a7-9271-4c06-944f-4ce340e70ab6-kube-api-access-xckx6\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:28 crc kubenswrapper[5035]: I1124 19:32:28.699850 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9599d9a7-9271-4c06-944f-4ce340e70ab6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.071014 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c8bf8bb97-lzjsb"] Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.099374 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-97bc6dcf7-zpd2w"] Nov 24 19:32:29 crc kubenswrapper[5035]: E1124 19:32:29.100024 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="961faa76-0c57-4ec1-891e-520dc13230bd" containerName="init" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.100041 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="961faa76-0c57-4ec1-891e-520dc13230bd" containerName="init" Nov 24 19:32:29 crc kubenswrapper[5035]: E1124 19:32:29.100051 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerName="init" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.100057 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerName="init" Nov 24 19:32:29 crc kubenswrapper[5035]: E1124 19:32:29.100093 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerName="dnsmasq-dns" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.100104 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerName="dnsmasq-dns" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.100267 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerName="dnsmasq-dns" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.100302 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="961faa76-0c57-4ec1-891e-520dc13230bd" containerName="init" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.101159 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.122339 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.134169 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-97bc6dcf7-zpd2w"] Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.153070 5035 generic.go:334] "Generic (PLEG): container finished" podID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerID="5b70d778c68167893f7706062c30d8e3ec0ca777807f27307e521175b2315545" exitCode=0 Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.153371 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" event={"ID":"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48","Type":"ContainerDied","Data":"5b70d778c68167893f7706062c30d8e3ec0ca777807f27307e521175b2315545"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.153415 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" event={"ID":"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48","Type":"ContainerStarted","Data":"8b331ffbbc33898aabf40c6b2d1443db8ece40e8d576a189fa2f7f4c46bb3fd3"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.162365 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h76mn" event={"ID":"764ce995-4314-4c29-8b00-dc264344edf0","Type":"ContainerStarted","Data":"d2970b231143622c6bdbc4f13c986140a5709797fcaef46757b2b3a1e4efe5ad"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.166017 5035 generic.go:334] "Generic (PLEG): container finished" podID="9599d9a7-9271-4c06-944f-4ce340e70ab6" containerID="272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a" exitCode=0 Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.166080 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" event={"ID":"9599d9a7-9271-4c06-944f-4ce340e70ab6","Type":"ContainerDied","Data":"272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.166107 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" event={"ID":"9599d9a7-9271-4c06-944f-4ce340e70ab6","Type":"ContainerDied","Data":"c237af6029c5f2c6b319c77600a479c1e43f1f487185a16ad3f800394ec64d92"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.166132 5035 scope.go:117] "RemoveContainer" containerID="272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.166273 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b58765b5-pvzps" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.183211 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xb7gt" event={"ID":"205bbe21-12fb-4cdf-bf2d-09008085f221","Type":"ContainerStarted","Data":"1eb02672bfc9ed20132b6b3b0be8dfc5c2e03ed66cde46f6ae4c784372d6d179"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.186564 5035 generic.go:334] "Generic (PLEG): container finished" podID="961faa76-0c57-4ec1-891e-520dc13230bd" containerID="b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242" exitCode=0 Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.187021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" event={"ID":"961faa76-0c57-4ec1-891e-520dc13230bd","Type":"ContainerDied","Data":"b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.187069 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" event={"ID":"961faa76-0c57-4ec1-891e-520dc13230bd","Type":"ContainerDied","Data":"0c22f86dd9b38ceaeefc917998c04bc4b01351565334ef582a1815b910b770d5"} Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.187135 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d44dbddd5-jgjfz" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.226958 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tstkp\" (UniqueName: \"kubernetes.io/projected/e7c1bf22-059d-45fc-9db2-1bcb791949bc-kube-api-access-tstkp\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.227066 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c1bf22-059d-45fc-9db2-1bcb791949bc-logs\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.227117 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-config-data\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.227141 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7c1bf22-059d-45fc-9db2-1bcb791949bc-horizon-secret-key\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.227174 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-scripts\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.290714 5035 scope.go:117] "RemoveContainer" containerID="f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.339889 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tstkp\" (UniqueName: \"kubernetes.io/projected/e7c1bf22-059d-45fc-9db2-1bcb791949bc-kube-api-access-tstkp\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.340139 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c1bf22-059d-45fc-9db2-1bcb791949bc-logs\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.340227 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-config-data\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.340270 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7c1bf22-059d-45fc-9db2-1bcb791949bc-horizon-secret-key\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.340340 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-scripts\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.341277 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-scripts\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.341441 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c1bf22-059d-45fc-9db2-1bcb791949bc-logs\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.342368 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-config-data\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.345788 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7c1bf22-059d-45fc-9db2-1bcb791949bc-horizon-secret-key\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.362588 5035 scope.go:117] "RemoveContainer" containerID="272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a" Nov 24 19:32:29 crc kubenswrapper[5035]: E1124 19:32:29.364472 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a\": container with ID starting with 272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a not found: ID does not exist" containerID="272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.364517 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a"} err="failed to get container status \"272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a\": rpc error: code = NotFound desc = could not find container \"272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a\": container with ID starting with 272d21a9126cd78a8206fe6fa1bd9228ff35b08314bc4e8d2ee41ed074bfdd6a not found: ID does not exist" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.364540 5035 scope.go:117] "RemoveContainer" containerID="f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86" Nov 24 19:32:29 crc kubenswrapper[5035]: E1124 19:32:29.367268 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86\": container with ID starting with f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86 not found: ID does not exist" containerID="f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.367322 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86"} err="failed to get container status \"f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86\": rpc error: code = NotFound desc = could not find container \"f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86\": container with ID starting with f14189d3f9d10a434438668b3143e9c9a42b9156854418306655a2bfec344a86 not found: ID does not exist" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.367346 5035 scope.go:117] "RemoveContainer" containerID="b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.375089 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75b58765b5-pvzps"] Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.383980 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75b58765b5-pvzps"] Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.388910 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tstkp\" (UniqueName: \"kubernetes.io/projected/e7c1bf22-059d-45fc-9db2-1bcb791949bc-kube-api-access-tstkp\") pod \"horizon-97bc6dcf7-zpd2w\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.427581 5035 scope.go:117] "RemoveContainer" containerID="b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242" Nov 24 19:32:29 crc kubenswrapper[5035]: E1124 19:32:29.431036 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242\": container with ID starting with b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242 not found: ID does not exist" containerID="b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.431067 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242"} err="failed to get container status \"b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242\": rpc error: code = NotFound desc = could not find container \"b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242\": container with ID starting with b59d3b37cde51ea423990bca41796e207fcb26d0fc185f493c7e54d9960c1242 not found: ID does not exist" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.434426 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.459548 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d44dbddd5-jgjfz"] Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.467487 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d44dbddd5-jgjfz"] Nov 24 19:32:29 crc kubenswrapper[5035]: I1124 19:32:29.930151 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-97bc6dcf7-zpd2w"] Nov 24 19:32:30 crc kubenswrapper[5035]: I1124 19:32:30.224812 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9599d9a7-9271-4c06-944f-4ce340e70ab6" path="/var/lib/kubelet/pods/9599d9a7-9271-4c06-944f-4ce340e70ab6/volumes" Nov 24 19:32:30 crc kubenswrapper[5035]: I1124 19:32:30.225530 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="961faa76-0c57-4ec1-891e-520dc13230bd" path="/var/lib/kubelet/pods/961faa76-0c57-4ec1-891e-520dc13230bd/volumes" Nov 24 19:32:30 crc kubenswrapper[5035]: I1124 19:32:30.226773 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:30 crc kubenswrapper[5035]: I1124 19:32:30.226807 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97bc6dcf7-zpd2w" event={"ID":"e7c1bf22-059d-45fc-9db2-1bcb791949bc","Type":"ContainerStarted","Data":"937d4283fa54c6ec32690a40cfb2c0c050e6b3b06e81eb901f17b582c9a907c2"} Nov 24 19:32:30 crc kubenswrapper[5035]: I1124 19:32:30.226830 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" event={"ID":"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48","Type":"ContainerStarted","Data":"c6764f8e05c3e71625b817583ec3c97fbccff709414175e646f4845b99af1ce6"} Nov 24 19:32:30 crc kubenswrapper[5035]: I1124 19:32:30.247268 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" podStartSLOduration=4.247251237 podStartE2EDuration="4.247251237s" podCreationTimestamp="2025-11-24 19:32:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:32:30.244348667 +0000 UTC m=+1088.766854934" watchObservedRunningTime="2025-11-24 19:32:30.247251237 +0000 UTC m=+1088.769757484" Nov 24 19:32:33 crc kubenswrapper[5035]: I1124 19:32:33.258882 5035 generic.go:334] "Generic (PLEG): container finished" podID="5bfe89ec-3b9c-4999-9955-d3769ce1caa2" containerID="f9d14da07ccc04044087429f74e3a615ac05903701b026bdc52ac7caa9d187e3" exitCode=0 Nov 24 19:32:33 crc kubenswrapper[5035]: I1124 19:32:33.259163 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sm9mf" event={"ID":"5bfe89ec-3b9c-4999-9955-d3769ce1caa2","Type":"ContainerDied","Data":"f9d14da07ccc04044087429f74e3a615ac05903701b026bdc52ac7caa9d187e3"} Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.641490 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.802882 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78mrs\" (UniqueName: \"kubernetes.io/projected/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-kube-api-access-78mrs\") pod \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.802972 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-fernet-keys\") pod \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.803100 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-scripts\") pod \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.803138 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-credential-keys\") pod \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.803171 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-combined-ca-bundle\") pod \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.803224 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-config-data\") pod \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\" (UID: \"5bfe89ec-3b9c-4999-9955-d3769ce1caa2\") " Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.822819 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-scripts" (OuterVolumeSpecName: "scripts") pod "5bfe89ec-3b9c-4999-9955-d3769ce1caa2" (UID: "5bfe89ec-3b9c-4999-9955-d3769ce1caa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.823134 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5bfe89ec-3b9c-4999-9955-d3769ce1caa2" (UID: "5bfe89ec-3b9c-4999-9955-d3769ce1caa2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.823247 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5bfe89ec-3b9c-4999-9955-d3769ce1caa2" (UID: "5bfe89ec-3b9c-4999-9955-d3769ce1caa2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.824059 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-kube-api-access-78mrs" (OuterVolumeSpecName: "kube-api-access-78mrs") pod "5bfe89ec-3b9c-4999-9955-d3769ce1caa2" (UID: "5bfe89ec-3b9c-4999-9955-d3769ce1caa2"). InnerVolumeSpecName "kube-api-access-78mrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.832426 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-config-data" (OuterVolumeSpecName: "config-data") pod "5bfe89ec-3b9c-4999-9955-d3769ce1caa2" (UID: "5bfe89ec-3b9c-4999-9955-d3769ce1caa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.838459 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bfe89ec-3b9c-4999-9955-d3769ce1caa2" (UID: "5bfe89ec-3b9c-4999-9955-d3769ce1caa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.905140 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.905174 5035 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.905185 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.905193 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.905203 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78mrs\" (UniqueName: \"kubernetes.io/projected/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-kube-api-access-78mrs\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:35 crc kubenswrapper[5035]: I1124 19:32:35.905211 5035 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5bfe89ec-3b9c-4999-9955-d3769ce1caa2-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.316744 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sm9mf" event={"ID":"5bfe89ec-3b9c-4999-9955-d3769ce1caa2","Type":"ContainerDied","Data":"c6500b660008ffbb72cf19abd7c43ae9f7cfb81bf5eec1e09e48607ec63f9f94"} Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.317081 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6500b660008ffbb72cf19abd7c43ae9f7cfb81bf5eec1e09e48607ec63f9f94" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.317027 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sm9mf" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.359794 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d46ff7b7c-jf8nx"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.384561 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6ff6c75f76-ttb9x"] Nov 24 19:32:36 crc kubenswrapper[5035]: E1124 19:32:36.384889 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfe89ec-3b9c-4999-9955-d3769ce1caa2" containerName="keystone-bootstrap" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.384901 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfe89ec-3b9c-4999-9955-d3769ce1caa2" containerName="keystone-bootstrap" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.385070 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfe89ec-3b9c-4999-9955-d3769ce1caa2" containerName="keystone-bootstrap" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.398199 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.401261 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.460249 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff6c75f76-ttb9x"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.495310 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-97bc6dcf7-zpd2w"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.503397 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-677648b7f8-j8p2d"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.504808 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.509725 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-677648b7f8-j8p2d"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.524722 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-secret-key\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.524760 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-config-data\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.524783 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e67a302e-93be-4341-a7e1-a52183398050-logs\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.524991 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-combined-ca-bundle\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.525089 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-tls-certs\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.525208 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqllc\" (UniqueName: \"kubernetes.io/projected/e67a302e-93be-4341-a7e1-a52183398050-kube-api-access-jqllc\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.525344 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-scripts\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626535 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-tls-certs\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626580 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c39b1fa-70b7-49f4-b318-0d08d5007dba-config-data\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626621 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96z22\" (UniqueName: \"kubernetes.io/projected/6c39b1fa-70b7-49f4-b318-0d08d5007dba-kube-api-access-96z22\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626664 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c39b1fa-70b7-49f4-b318-0d08d5007dba-scripts\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626684 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-horizon-secret-key\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626703 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqllc\" (UniqueName: \"kubernetes.io/projected/e67a302e-93be-4341-a7e1-a52183398050-kube-api-access-jqllc\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626733 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c39b1fa-70b7-49f4-b318-0d08d5007dba-logs\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626773 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-scripts\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626804 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-secret-key\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626828 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-combined-ca-bundle\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626869 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-config-data\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.626904 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e67a302e-93be-4341-a7e1-a52183398050-logs\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.627235 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-combined-ca-bundle\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.627309 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-horizon-tls-certs\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.629058 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e67a302e-93be-4341-a7e1-a52183398050-logs\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.629418 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-scripts\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.629612 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-config-data\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.631233 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-tls-certs\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.631345 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-secret-key\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.632513 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-combined-ca-bundle\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.644269 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqllc\" (UniqueName: \"kubernetes.io/projected/e67a302e-93be-4341-a7e1-a52183398050-kube-api-access-jqllc\") pod \"horizon-6ff6c75f76-ttb9x\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.722259 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.729062 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c39b1fa-70b7-49f4-b318-0d08d5007dba-logs\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.729163 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-combined-ca-bundle\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.730045 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c39b1fa-70b7-49f4-b318-0d08d5007dba-logs\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.730993 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-horizon-tls-certs\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.731673 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c39b1fa-70b7-49f4-b318-0d08d5007dba-config-data\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.731790 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96z22\" (UniqueName: \"kubernetes.io/projected/6c39b1fa-70b7-49f4-b318-0d08d5007dba-kube-api-access-96z22\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.731845 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c39b1fa-70b7-49f4-b318-0d08d5007dba-scripts\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.731876 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-horizon-secret-key\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.733037 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c39b1fa-70b7-49f4-b318-0d08d5007dba-scripts\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.733667 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c39b1fa-70b7-49f4-b318-0d08d5007dba-config-data\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.735779 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-horizon-secret-key\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.736011 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-combined-ca-bundle\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.736013 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c39b1fa-70b7-49f4-b318-0d08d5007dba-horizon-tls-certs\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.750816 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96z22\" (UniqueName: \"kubernetes.io/projected/6c39b1fa-70b7-49f4-b318-0d08d5007dba-kube-api-access-96z22\") pod \"horizon-677648b7f8-j8p2d\" (UID: \"6c39b1fa-70b7-49f4-b318-0d08d5007dba\") " pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.775843 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sm9mf"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.782754 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sm9mf"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.824987 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.867398 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kfcrg"] Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.868652 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.872055 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.872368 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.872479 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.872605 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2trk2" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.872665 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 19:32:36 crc kubenswrapper[5035]: I1124 19:32:36.875276 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kfcrg"] Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.039384 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-combined-ca-bundle\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.039477 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-fernet-keys\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.039501 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f486j\" (UniqueName: \"kubernetes.io/projected/391f762a-cd72-4011-abc8-6efc3fce4661-kube-api-access-f486j\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.039562 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-credential-keys\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.039644 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-config-data\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.039691 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-scripts\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.141193 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-combined-ca-bundle\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.141269 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-fernet-keys\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.141311 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f486j\" (UniqueName: \"kubernetes.io/projected/391f762a-cd72-4011-abc8-6efc3fce4661-kube-api-access-f486j\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.141364 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-credential-keys\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.141451 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-config-data\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.141493 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-scripts\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.145611 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-config-data\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.147056 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-scripts\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.147358 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-combined-ca-bundle\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.148074 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-credential-keys\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.151787 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-fernet-keys\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.170980 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f486j\" (UniqueName: \"kubernetes.io/projected/391f762a-cd72-4011-abc8-6efc3fce4661-kube-api-access-f486j\") pod \"keystone-bootstrap-kfcrg\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.189937 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.427172 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.506433 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-7vcb7"] Nov 24 19:32:37 crc kubenswrapper[5035]: I1124 19:32:37.506666 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="dnsmasq-dns" containerID="cri-o://0e9ed067c749b72e78ae141bde292b7b156ab3d786b8866e1a565dfd19a5ce7c" gracePeriod=10 Nov 24 19:32:37 crc kubenswrapper[5035]: E1124 19:32:37.590384 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4875d758_6b35_47c1_8b60_80a8d718fdbd.slice/crio-conmon-0e9ed067c749b72e78ae141bde292b7b156ab3d786b8866e1a565dfd19a5ce7c.scope\": RecentStats: unable to find data in memory cache]" Nov 24 19:32:38 crc kubenswrapper[5035]: I1124 19:32:38.209835 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bfe89ec-3b9c-4999-9955-d3769ce1caa2" path="/var/lib/kubelet/pods/5bfe89ec-3b9c-4999-9955-d3769ce1caa2/volumes" Nov 24 19:32:38 crc kubenswrapper[5035]: I1124 19:32:38.377507 5035 generic.go:334] "Generic (PLEG): container finished" podID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerID="0e9ed067c749b72e78ae141bde292b7b156ab3d786b8866e1a565dfd19a5ce7c" exitCode=0 Nov 24 19:32:38 crc kubenswrapper[5035]: I1124 19:32:38.377548 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" event={"ID":"4875d758-6b35-47c1-8b60-80a8d718fdbd","Type":"ContainerDied","Data":"0e9ed067c749b72e78ae141bde292b7b156ab3d786b8866e1a565dfd19a5ce7c"} Nov 24 19:32:39 crc kubenswrapper[5035]: I1124 19:32:39.381812 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Nov 24 19:32:42 crc kubenswrapper[5035]: E1124 19:32:42.042921 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057" Nov 24 19:32:42 crc kubenswrapper[5035]: E1124 19:32:42.043401 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n86h55h5f5h5c9hbh665h5b8h5cdh676h556hd6hb4h5fch68chf9hdbh5b7h5h56dh557hd8h678h54dh564h7ch74h554h5d4h556h5d9h5b7h554q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zfzv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7c8bf8bb97-lzjsb_openstack(f6391e26-277d-4d6e-bbf1-82046cbfe80f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:32:42 crc kubenswrapper[5035]: E1124 19:32:42.045807 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057\\\"\"]" pod="openstack/horizon-7c8bf8bb97-lzjsb" podUID="f6391e26-277d-4d6e-bbf1-82046cbfe80f" Nov 24 19:32:42 crc kubenswrapper[5035]: E1124 19:32:42.473480 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057" Nov 24 19:32:42 crc kubenswrapper[5035]: E1124 19:32:42.473913 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5ddh558h654hfh5dbh54bh55ch5d8h659h5bh578h94hdfh67h5cbh5f5h5f7h58ch58dh5b9h666hb7h576hcch59bhcfh57h66dh57fh665h68bhd5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s7www,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-d46ff7b7c-jf8nx_openstack(aae039f5-86cf-46b9-a9c4-c3a18849bfec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:32:42 crc kubenswrapper[5035]: E1124 19:32:42.477350 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057\\\"\"]" pod="openstack/horizon-d46ff7b7c-jf8nx" podUID="aae039f5-86cf-46b9-a9c4-c3a18849bfec" Nov 24 19:32:43 crc kubenswrapper[5035]: E1124 19:32:43.569593 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099" Nov 24 19:32:43 crc kubenswrapper[5035]: E1124 19:32:43.570152 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mvlqp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-h76mn_openstack(764ce995-4314-4c29-8b00-dc264344edf0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:32:43 crc kubenswrapper[5035]: E1124 19:32:43.572499 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-h76mn" podUID="764ce995-4314-4c29-8b00-dc264344edf0" Nov 24 19:32:44 crc kubenswrapper[5035]: E1124 19:32:44.184783 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140" Nov 24 19:32:44 crc kubenswrapper[5035]: E1124 19:32:44.185231 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n58fh5fh58h668h577hdfh694hb6h56dh54h6bhd8h58fh599hd9h5chbh599h5d8h66h597h97h686h644h5dfh5bdhf8h5c5h5d4h5fdh94h5d6q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k7md9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(f9739bc1-965c-4f67-9b14-b96e176d72c1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:32:44 crc kubenswrapper[5035]: I1124 19:32:44.381655 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Nov 24 19:32:44 crc kubenswrapper[5035]: E1124 19:32:44.433567 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099\\\"\"" pod="openstack/placement-db-sync-h76mn" podUID="764ce995-4314-4c29-8b00-dc264344edf0" Nov 24 19:32:44 crc kubenswrapper[5035]: E1124 19:32:44.828150 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645" Nov 24 19:32:44 crc kubenswrapper[5035]: E1124 19:32:44.828301 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ckjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-xb7gt_openstack(205bbe21-12fb-4cdf-bf2d-09008085f221): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:32:44 crc kubenswrapper[5035]: E1124 19:32:44.829542 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-xb7gt" podUID="205bbe21-12fb-4cdf-bf2d-09008085f221" Nov 24 19:32:45 crc kubenswrapper[5035]: I1124 19:32:45.234166 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:32:45 crc kubenswrapper[5035]: I1124 19:32:45.234554 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:32:45 crc kubenswrapper[5035]: E1124 19:32:45.445683 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645\\\"\"" pod="openstack/barbican-db-sync-xb7gt" podUID="205bbe21-12fb-4cdf-bf2d-09008085f221" Nov 24 19:32:49 crc kubenswrapper[5035]: I1124 19:32:49.381483 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Nov 24 19:32:49 crc kubenswrapper[5035]: I1124 19:32:49.381921 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.473486 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.490204 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.511947 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfzv5\" (UniqueName: \"kubernetes.io/projected/f6391e26-277d-4d6e-bbf1-82046cbfe80f-kube-api-access-zfzv5\") pod \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.512069 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6391e26-277d-4d6e-bbf1-82046cbfe80f-logs\") pod \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.512141 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-scripts\") pod \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.512254 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-config-data\") pod \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.512769 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6391e26-277d-4d6e-bbf1-82046cbfe80f-logs" (OuterVolumeSpecName: "logs") pod "f6391e26-277d-4d6e-bbf1-82046cbfe80f" (UID: "f6391e26-277d-4d6e-bbf1-82046cbfe80f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.512819 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f6391e26-277d-4d6e-bbf1-82046cbfe80f-horizon-secret-key\") pod \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\" (UID: \"f6391e26-277d-4d6e-bbf1-82046cbfe80f\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.512909 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-scripts" (OuterVolumeSpecName: "scripts") pod "f6391e26-277d-4d6e-bbf1-82046cbfe80f" (UID: "f6391e26-277d-4d6e-bbf1-82046cbfe80f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.513332 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6391e26-277d-4d6e-bbf1-82046cbfe80f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.513354 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.513920 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-config-data" (OuterVolumeSpecName: "config-data") pod "f6391e26-277d-4d6e-bbf1-82046cbfe80f" (UID: "f6391e26-277d-4d6e-bbf1-82046cbfe80f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.514151 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d46ff7b7c-jf8nx" event={"ID":"aae039f5-86cf-46b9-a9c4-c3a18849bfec","Type":"ContainerDied","Data":"473944120fc3108dfc9deb80eda5e95ce087929ac2d5a8df9186602df8bcb910"} Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.514251 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d46ff7b7c-jf8nx" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.516637 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c8bf8bb97-lzjsb" event={"ID":"f6391e26-277d-4d6e-bbf1-82046cbfe80f","Type":"ContainerDied","Data":"3616bf401bb3c4a536edf5411bc9ab95375cd152ef62267aa73f3e0a99fe0125"} Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.516710 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c8bf8bb97-lzjsb" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.520015 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6391e26-277d-4d6e-bbf1-82046cbfe80f-kube-api-access-zfzv5" (OuterVolumeSpecName: "kube-api-access-zfzv5") pod "f6391e26-277d-4d6e-bbf1-82046cbfe80f" (UID: "f6391e26-277d-4d6e-bbf1-82046cbfe80f"). InnerVolumeSpecName "kube-api-access-zfzv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.524238 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6391e26-277d-4d6e-bbf1-82046cbfe80f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f6391e26-277d-4d6e-bbf1-82046cbfe80f" (UID: "f6391e26-277d-4d6e-bbf1-82046cbfe80f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.614440 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7www\" (UniqueName: \"kubernetes.io/projected/aae039f5-86cf-46b9-a9c4-c3a18849bfec-kube-api-access-s7www\") pod \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.614648 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aae039f5-86cf-46b9-a9c4-c3a18849bfec-horizon-secret-key\") pod \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.614735 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-config-data\") pod \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.614803 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae039f5-86cf-46b9-a9c4-c3a18849bfec-logs\") pod \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.614823 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-scripts\") pod \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\" (UID: \"aae039f5-86cf-46b9-a9c4-c3a18849bfec\") " Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.615267 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6391e26-277d-4d6e-bbf1-82046cbfe80f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.615282 5035 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f6391e26-277d-4d6e-bbf1-82046cbfe80f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.615311 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfzv5\" (UniqueName: \"kubernetes.io/projected/f6391e26-277d-4d6e-bbf1-82046cbfe80f-kube-api-access-zfzv5\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.615619 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae039f5-86cf-46b9-a9c4-c3a18849bfec-logs" (OuterVolumeSpecName: "logs") pod "aae039f5-86cf-46b9-a9c4-c3a18849bfec" (UID: "aae039f5-86cf-46b9-a9c4-c3a18849bfec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.615688 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-scripts" (OuterVolumeSpecName: "scripts") pod "aae039f5-86cf-46b9-a9c4-c3a18849bfec" (UID: "aae039f5-86cf-46b9-a9c4-c3a18849bfec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.616234 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-config-data" (OuterVolumeSpecName: "config-data") pod "aae039f5-86cf-46b9-a9c4-c3a18849bfec" (UID: "aae039f5-86cf-46b9-a9c4-c3a18849bfec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.621537 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae039f5-86cf-46b9-a9c4-c3a18849bfec-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "aae039f5-86cf-46b9-a9c4-c3a18849bfec" (UID: "aae039f5-86cf-46b9-a9c4-c3a18849bfec"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.621866 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae039f5-86cf-46b9-a9c4-c3a18849bfec-kube-api-access-s7www" (OuterVolumeSpecName: "kube-api-access-s7www") pod "aae039f5-86cf-46b9-a9c4-c3a18849bfec" (UID: "aae039f5-86cf-46b9-a9c4-c3a18849bfec"). InnerVolumeSpecName "kube-api-access-s7www". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.716695 5035 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aae039f5-86cf-46b9-a9c4-c3a18849bfec-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.716731 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.716745 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aae039f5-86cf-46b9-a9c4-c3a18849bfec-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.716756 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aae039f5-86cf-46b9-a9c4-c3a18849bfec-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.716769 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7www\" (UniqueName: \"kubernetes.io/projected/aae039f5-86cf-46b9-a9c4-c3a18849bfec-kube-api-access-s7www\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.880345 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c8bf8bb97-lzjsb"] Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.885315 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c8bf8bb97-lzjsb"] Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.913780 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d46ff7b7c-jf8nx"] Nov 24 19:32:52 crc kubenswrapper[5035]: I1124 19:32:52.924868 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d46ff7b7c-jf8nx"] Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.076114 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:32:54 crc kubenswrapper[5035]: E1124 19:32:54.139870 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879" Nov 24 19:32:54 crc kubenswrapper[5035]: E1124 19:32:54.140028 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dl5n6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-nbbjc_openstack(0c9794ab-e4e3-4af0-a689-6ef761d1fe4f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 19:32:54 crc kubenswrapper[5035]: E1124 19:32:54.141330 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-nbbjc" podUID="0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.149154 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-nb\") pod \"4875d758-6b35-47c1-8b60-80a8d718fdbd\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.149248 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-sb\") pod \"4875d758-6b35-47c1-8b60-80a8d718fdbd\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.149316 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-config\") pod \"4875d758-6b35-47c1-8b60-80a8d718fdbd\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.149348 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwgxf\" (UniqueName: \"kubernetes.io/projected/4875d758-6b35-47c1-8b60-80a8d718fdbd-kube-api-access-gwgxf\") pod \"4875d758-6b35-47c1-8b60-80a8d718fdbd\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.149461 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-dns-svc\") pod \"4875d758-6b35-47c1-8b60-80a8d718fdbd\" (UID: \"4875d758-6b35-47c1-8b60-80a8d718fdbd\") " Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.168562 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4875d758-6b35-47c1-8b60-80a8d718fdbd-kube-api-access-gwgxf" (OuterVolumeSpecName: "kube-api-access-gwgxf") pod "4875d758-6b35-47c1-8b60-80a8d718fdbd" (UID: "4875d758-6b35-47c1-8b60-80a8d718fdbd"). InnerVolumeSpecName "kube-api-access-gwgxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.206490 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4875d758-6b35-47c1-8b60-80a8d718fdbd" (UID: "4875d758-6b35-47c1-8b60-80a8d718fdbd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.208831 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae039f5-86cf-46b9-a9c4-c3a18849bfec" path="/var/lib/kubelet/pods/aae039f5-86cf-46b9-a9c4-c3a18849bfec/volumes" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.209395 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6391e26-277d-4d6e-bbf1-82046cbfe80f" path="/var/lib/kubelet/pods/f6391e26-277d-4d6e-bbf1-82046cbfe80f/volumes" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.214958 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4875d758-6b35-47c1-8b60-80a8d718fdbd" (UID: "4875d758-6b35-47c1-8b60-80a8d718fdbd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.216750 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-config" (OuterVolumeSpecName: "config") pod "4875d758-6b35-47c1-8b60-80a8d718fdbd" (UID: "4875d758-6b35-47c1-8b60-80a8d718fdbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.239649 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4875d758-6b35-47c1-8b60-80a8d718fdbd" (UID: "4875d758-6b35-47c1-8b60-80a8d718fdbd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.251197 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.251229 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.251239 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.251249 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwgxf\" (UniqueName: \"kubernetes.io/projected/4875d758-6b35-47c1-8b60-80a8d718fdbd-kube-api-access-gwgxf\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.251258 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4875d758-6b35-47c1-8b60-80a8d718fdbd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.328355 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kfcrg"] Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.453248 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-677648b7f8-j8p2d"] Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.470605 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff6c75f76-ttb9x"] Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.537828 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" event={"ID":"4875d758-6b35-47c1-8b60-80a8d718fdbd","Type":"ContainerDied","Data":"1e154e6a4e2d4a3aa5d2bc56baded04402ee48c3d4cc46a068cbbe7038235e3f"} Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.537871 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-7vcb7" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.537905 5035 scope.go:117] "RemoveContainer" containerID="0e9ed067c749b72e78ae141bde292b7b156ab3d786b8866e1a565dfd19a5ce7c" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.542236 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97bc6dcf7-zpd2w" event={"ID":"e7c1bf22-059d-45fc-9db2-1bcb791949bc","Type":"ContainerStarted","Data":"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7"} Nov 24 19:32:54 crc kubenswrapper[5035]: E1124 19:32:54.545089 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879\\\"\"" pod="openstack/cinder-db-sync-nbbjc" podUID="0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.580583 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-7vcb7"] Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.586794 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-7vcb7"] Nov 24 19:32:54 crc kubenswrapper[5035]: W1124 19:32:54.756462 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod391f762a_cd72_4011_abc8_6efc3fce4661.slice/crio-a98bd4416aae375fd104ecf1b858d94bcbf42e2f65dc0764556b2f934636ca36 WatchSource:0}: Error finding container a98bd4416aae375fd104ecf1b858d94bcbf42e2f65dc0764556b2f934636ca36: Status 404 returned error can't find the container with id a98bd4416aae375fd104ecf1b858d94bcbf42e2f65dc0764556b2f934636ca36 Nov 24 19:32:54 crc kubenswrapper[5035]: W1124 19:32:54.759335 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode67a302e_93be_4341_a7e1_a52183398050.slice/crio-174a679c2a34a3ba9feea167526f1c5a5b875edfe4b94d83ee1f93abd7389b11 WatchSource:0}: Error finding container 174a679c2a34a3ba9feea167526f1c5a5b875edfe4b94d83ee1f93abd7389b11: Status 404 returned error can't find the container with id 174a679c2a34a3ba9feea167526f1c5a5b875edfe4b94d83ee1f93abd7389b11 Nov 24 19:32:54 crc kubenswrapper[5035]: I1124 19:32:54.770529 5035 scope.go:117] "RemoveContainer" containerID="0a3e0fc8da37a85db5fb228b02cc23c13ba574252793f3febca5965fcd702f2a" Nov 24 19:32:55 crc kubenswrapper[5035]: I1124 19:32:55.555395 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff6c75f76-ttb9x" event={"ID":"e67a302e-93be-4341-a7e1-a52183398050","Type":"ContainerStarted","Data":"df984fe11df0c4b1f22de9b25a3845a96fb7f38e7c96e121c2987ad75c805c5c"} Nov 24 19:32:55 crc kubenswrapper[5035]: I1124 19:32:55.555727 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff6c75f76-ttb9x" event={"ID":"e67a302e-93be-4341-a7e1-a52183398050","Type":"ContainerStarted","Data":"174a679c2a34a3ba9feea167526f1c5a5b875edfe4b94d83ee1f93abd7389b11"} Nov 24 19:32:55 crc kubenswrapper[5035]: I1124 19:32:55.559419 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97bc6dcf7-zpd2w" event={"ID":"e7c1bf22-059d-45fc-9db2-1bcb791949bc","Type":"ContainerStarted","Data":"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a"} Nov 24 19:32:55 crc kubenswrapper[5035]: I1124 19:32:55.560970 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kfcrg" event={"ID":"391f762a-cd72-4011-abc8-6efc3fce4661","Type":"ContainerStarted","Data":"2ee8fc07fd677e9f8fc0399a5dab162e4cd2870a2b6cf1bf35ff520a6563ad3f"} Nov 24 19:32:55 crc kubenswrapper[5035]: I1124 19:32:55.561005 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kfcrg" event={"ID":"391f762a-cd72-4011-abc8-6efc3fce4661","Type":"ContainerStarted","Data":"a98bd4416aae375fd104ecf1b858d94bcbf42e2f65dc0764556b2f934636ca36"} Nov 24 19:32:55 crc kubenswrapper[5035]: I1124 19:32:55.562354 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-677648b7f8-j8p2d" event={"ID":"6c39b1fa-70b7-49f4-b318-0d08d5007dba","Type":"ContainerStarted","Data":"741bcc91cde9e2eca03bf9ae13fe7a212765ae659c9250f4f2a1a40ae53c6965"} Nov 24 19:32:55 crc kubenswrapper[5035]: I1124 19:32:55.562448 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-677648b7f8-j8p2d" event={"ID":"6c39b1fa-70b7-49f4-b318-0d08d5007dba","Type":"ContainerStarted","Data":"f116f93264d2683f7385adf0e9af124ca71b7ef65a287c17f1904dd783c24433"} Nov 24 19:32:56 crc kubenswrapper[5035]: I1124 19:32:56.215429 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" path="/var/lib/kubelet/pods/4875d758-6b35-47c1-8b60-80a8d718fdbd/volumes" Nov 24 19:32:56 crc kubenswrapper[5035]: I1124 19:32:56.570390 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-97bc6dcf7-zpd2w" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon-log" containerID="cri-o://0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7" gracePeriod=30 Nov 24 19:32:56 crc kubenswrapper[5035]: I1124 19:32:56.570508 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-97bc6dcf7-zpd2w" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon" containerID="cri-o://50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a" gracePeriod=30 Nov 24 19:32:56 crc kubenswrapper[5035]: I1124 19:32:56.597143 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-97bc6dcf7-zpd2w" podStartSLOduration=3.645695638 podStartE2EDuration="27.597121605s" podCreationTimestamp="2025-11-24 19:32:29 +0000 UTC" firstStartedPulling="2025-11-24 19:32:29.95488463 +0000 UTC m=+1088.477390887" lastFinishedPulling="2025-11-24 19:32:53.906310597 +0000 UTC m=+1112.428816854" observedRunningTime="2025-11-24 19:32:56.593852886 +0000 UTC m=+1115.116359143" watchObservedRunningTime="2025-11-24 19:32:56.597121605 +0000 UTC m=+1115.119627862" Nov 24 19:33:01 crc kubenswrapper[5035]: I1124 19:32:59.435088 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:33:01 crc kubenswrapper[5035]: I1124 19:33:00.606751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff6c75f76-ttb9x" event={"ID":"e67a302e-93be-4341-a7e1-a52183398050","Type":"ContainerStarted","Data":"b273b803cecfebabd01ad70881af43d61fe52dfee628924d6091e2d362e9ca4c"} Nov 24 19:33:01 crc kubenswrapper[5035]: I1124 19:33:00.637487 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6ff6c75f76-ttb9x" podStartSLOduration=24.637461707 podStartE2EDuration="24.637461707s" podCreationTimestamp="2025-11-24 19:32:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:00.629651135 +0000 UTC m=+1119.152157392" watchObservedRunningTime="2025-11-24 19:33:00.637461707 +0000 UTC m=+1119.159967974" Nov 24 19:33:01 crc kubenswrapper[5035]: I1124 19:33:01.618228 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-677648b7f8-j8p2d" event={"ID":"6c39b1fa-70b7-49f4-b318-0d08d5007dba","Type":"ContainerStarted","Data":"fe841be8c9656a2e5e0ba31c840da257aa40e0856bff2bec6b41480cf900a63f"} Nov 24 19:33:01 crc kubenswrapper[5035]: I1124 19:33:01.665931 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-677648b7f8-j8p2d" podStartSLOduration=25.665896133 podStartE2EDuration="25.665896133s" podCreationTimestamp="2025-11-24 19:32:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:01.660023234 +0000 UTC m=+1120.182529491" watchObservedRunningTime="2025-11-24 19:33:01.665896133 +0000 UTC m=+1120.188402390" Nov 24 19:33:01 crc kubenswrapper[5035]: I1124 19:33:01.669338 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kfcrg" podStartSLOduration=25.669320007 podStartE2EDuration="25.669320007s" podCreationTimestamp="2025-11-24 19:32:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:01.634876971 +0000 UTC m=+1120.157383228" watchObservedRunningTime="2025-11-24 19:33:01.669320007 +0000 UTC m=+1120.191826274" Nov 24 19:33:02 crc kubenswrapper[5035]: I1124 19:33:02.625661 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerStarted","Data":"8040e70ce1911937f52f82371500b42c5ac5ce93dfae37ee13dfafc26794c317"} Nov 24 19:33:02 crc kubenswrapper[5035]: I1124 19:33:02.627610 5035 generic.go:334] "Generic (PLEG): container finished" podID="391f762a-cd72-4011-abc8-6efc3fce4661" containerID="2ee8fc07fd677e9f8fc0399a5dab162e4cd2870a2b6cf1bf35ff520a6563ad3f" exitCode=0 Nov 24 19:33:02 crc kubenswrapper[5035]: I1124 19:33:02.627654 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kfcrg" event={"ID":"391f762a-cd72-4011-abc8-6efc3fce4661","Type":"ContainerDied","Data":"2ee8fc07fd677e9f8fc0399a5dab162e4cd2870a2b6cf1bf35ff520a6563ad3f"} Nov 24 19:33:02 crc kubenswrapper[5035]: I1124 19:33:02.631531 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xb7gt" event={"ID":"205bbe21-12fb-4cdf-bf2d-09008085f221","Type":"ContainerStarted","Data":"1ae6fe50bc26642ebd076d87c538a9201743427dfe2b961d2e5836739ba4d120"} Nov 24 19:33:02 crc kubenswrapper[5035]: I1124 19:33:02.634070 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h76mn" event={"ID":"764ce995-4314-4c29-8b00-dc264344edf0","Type":"ContainerStarted","Data":"700413af8825bd3f07de635668905592f6f45b1f467cfbebf94e1ef465dfacd9"} Nov 24 19:33:02 crc kubenswrapper[5035]: I1124 19:33:02.671206 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-h76mn" podStartSLOduration=3.320047238 podStartE2EDuration="36.671184481s" podCreationTimestamp="2025-11-24 19:32:26 +0000 UTC" firstStartedPulling="2025-11-24 19:32:28.119439437 +0000 UTC m=+1086.641945694" lastFinishedPulling="2025-11-24 19:33:01.47057668 +0000 UTC m=+1119.993082937" observedRunningTime="2025-11-24 19:33:02.664581872 +0000 UTC m=+1121.187088119" watchObservedRunningTime="2025-11-24 19:33:02.671184481 +0000 UTC m=+1121.193690738" Nov 24 19:33:02 crc kubenswrapper[5035]: I1124 19:33:02.686277 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xb7gt" podStartSLOduration=3.416136537 podStartE2EDuration="36.686257611s" podCreationTimestamp="2025-11-24 19:32:26 +0000 UTC" firstStartedPulling="2025-11-24 19:32:28.199076568 +0000 UTC m=+1086.721582825" lastFinishedPulling="2025-11-24 19:33:01.469197652 +0000 UTC m=+1119.991703899" observedRunningTime="2025-11-24 19:33:02.681179482 +0000 UTC m=+1121.203685739" watchObservedRunningTime="2025-11-24 19:33:02.686257611 +0000 UTC m=+1121.208763868" Nov 24 19:33:06 crc kubenswrapper[5035]: I1124 19:33:06.722860 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:33:06 crc kubenswrapper[5035]: I1124 19:33:06.723494 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:33:06 crc kubenswrapper[5035]: I1124 19:33:06.825833 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:33:06 crc kubenswrapper[5035]: I1124 19:33:06.825877 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.142073 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.206644 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-fernet-keys\") pod \"391f762a-cd72-4011-abc8-6efc3fce4661\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.206776 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f486j\" (UniqueName: \"kubernetes.io/projected/391f762a-cd72-4011-abc8-6efc3fce4661-kube-api-access-f486j\") pod \"391f762a-cd72-4011-abc8-6efc3fce4661\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.206812 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-config-data\") pod \"391f762a-cd72-4011-abc8-6efc3fce4661\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.206903 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-credential-keys\") pod \"391f762a-cd72-4011-abc8-6efc3fce4661\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.207071 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-scripts\") pod \"391f762a-cd72-4011-abc8-6efc3fce4661\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.207131 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-combined-ca-bundle\") pod \"391f762a-cd72-4011-abc8-6efc3fce4661\" (UID: \"391f762a-cd72-4011-abc8-6efc3fce4661\") " Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.216894 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "391f762a-cd72-4011-abc8-6efc3fce4661" (UID: "391f762a-cd72-4011-abc8-6efc3fce4661"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.219778 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "391f762a-cd72-4011-abc8-6efc3fce4661" (UID: "391f762a-cd72-4011-abc8-6efc3fce4661"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.225018 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-scripts" (OuterVolumeSpecName: "scripts") pod "391f762a-cd72-4011-abc8-6efc3fce4661" (UID: "391f762a-cd72-4011-abc8-6efc3fce4661"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.225133 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/391f762a-cd72-4011-abc8-6efc3fce4661-kube-api-access-f486j" (OuterVolumeSpecName: "kube-api-access-f486j") pod "391f762a-cd72-4011-abc8-6efc3fce4661" (UID: "391f762a-cd72-4011-abc8-6efc3fce4661"). InnerVolumeSpecName "kube-api-access-f486j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.243105 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "391f762a-cd72-4011-abc8-6efc3fce4661" (UID: "391f762a-cd72-4011-abc8-6efc3fce4661"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.255408 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-config-data" (OuterVolumeSpecName: "config-data") pod "391f762a-cd72-4011-abc8-6efc3fce4661" (UID: "391f762a-cd72-4011-abc8-6efc3fce4661"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.309433 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.309722 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.309734 5035 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.309745 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f486j\" (UniqueName: \"kubernetes.io/projected/391f762a-cd72-4011-abc8-6efc3fce4661-kube-api-access-f486j\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.309753 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.309761 5035 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/391f762a-cd72-4011-abc8-6efc3fce4661-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.683497 5035 generic.go:334] "Generic (PLEG): container finished" podID="764ce995-4314-4c29-8b00-dc264344edf0" containerID="700413af8825bd3f07de635668905592f6f45b1f467cfbebf94e1ef465dfacd9" exitCode=0 Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.683566 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h76mn" event={"ID":"764ce995-4314-4c29-8b00-dc264344edf0","Type":"ContainerDied","Data":"700413af8825bd3f07de635668905592f6f45b1f467cfbebf94e1ef465dfacd9"} Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.687797 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerStarted","Data":"101366c589974592e10246af0ea4216583154d7bf940f6637349259623ae22c3"} Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.700780 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kfcrg" event={"ID":"391f762a-cd72-4011-abc8-6efc3fce4661","Type":"ContainerDied","Data":"a98bd4416aae375fd104ecf1b858d94bcbf42e2f65dc0764556b2f934636ca36"} Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.700861 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a98bd4416aae375fd104ecf1b858d94bcbf42e2f65dc0764556b2f934636ca36" Nov 24 19:33:07 crc kubenswrapper[5035]: I1124 19:33:07.700962 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kfcrg" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.252044 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c9499d9cb-dtnz8"] Nov 24 19:33:08 crc kubenswrapper[5035]: E1124 19:33:08.253928 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="dnsmasq-dns" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.254812 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="dnsmasq-dns" Nov 24 19:33:08 crc kubenswrapper[5035]: E1124 19:33:08.254907 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="init" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.257193 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="init" Nov 24 19:33:08 crc kubenswrapper[5035]: E1124 19:33:08.257384 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391f762a-cd72-4011-abc8-6efc3fce4661" containerName="keystone-bootstrap" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.257457 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="391f762a-cd72-4011-abc8-6efc3fce4661" containerName="keystone-bootstrap" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.257820 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="391f762a-cd72-4011-abc8-6efc3fce4661" containerName="keystone-bootstrap" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.257949 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4875d758-6b35-47c1-8b60-80a8d718fdbd" containerName="dnsmasq-dns" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.258768 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.262100 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2trk2" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.262341 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.262468 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.262685 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.262833 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.263110 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.271403 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c9499d9cb-dtnz8"] Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327313 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-combined-ca-bundle\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327393 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-credential-keys\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327413 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-internal-tls-certs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327469 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhgjs\" (UniqueName: \"kubernetes.io/projected/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-kube-api-access-qhgjs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327510 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-fernet-keys\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327549 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-public-tls-certs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327596 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-scripts\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.327626 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-config-data\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429215 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-credential-keys\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-internal-tls-certs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429339 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhgjs\" (UniqueName: \"kubernetes.io/projected/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-kube-api-access-qhgjs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429386 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-fernet-keys\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429429 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-public-tls-certs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429467 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-scripts\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429497 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-config-data\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.429554 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-combined-ca-bundle\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.435795 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-combined-ca-bundle\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.436299 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-config-data\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.437979 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-public-tls-certs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.437978 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-internal-tls-certs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.440603 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-scripts\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.441599 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-credential-keys\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.449970 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhgjs\" (UniqueName: \"kubernetes.io/projected/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-kube-api-access-qhgjs\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.459652 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae-fernet-keys\") pod \"keystone-c9499d9cb-dtnz8\" (UID: \"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae\") " pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.622715 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.721098 5035 generic.go:334] "Generic (PLEG): container finished" podID="205bbe21-12fb-4cdf-bf2d-09008085f221" containerID="1ae6fe50bc26642ebd076d87c538a9201743427dfe2b961d2e5836739ba4d120" exitCode=0 Nov 24 19:33:08 crc kubenswrapper[5035]: I1124 19:33:08.721590 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xb7gt" event={"ID":"205bbe21-12fb-4cdf-bf2d-09008085f221","Type":"ContainerDied","Data":"1ae6fe50bc26642ebd076d87c538a9201743427dfe2b961d2e5836739ba4d120"} Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.012241 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h76mn" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.087174 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c9499d9cb-dtnz8"] Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.144408 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-combined-ca-bundle\") pod \"764ce995-4314-4c29-8b00-dc264344edf0\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.144457 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvlqp\" (UniqueName: \"kubernetes.io/projected/764ce995-4314-4c29-8b00-dc264344edf0-kube-api-access-mvlqp\") pod \"764ce995-4314-4c29-8b00-dc264344edf0\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.144534 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-scripts\") pod \"764ce995-4314-4c29-8b00-dc264344edf0\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.144656 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-config-data\") pod \"764ce995-4314-4c29-8b00-dc264344edf0\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.144689 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/764ce995-4314-4c29-8b00-dc264344edf0-logs\") pod \"764ce995-4314-4c29-8b00-dc264344edf0\" (UID: \"764ce995-4314-4c29-8b00-dc264344edf0\") " Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.145431 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/764ce995-4314-4c29-8b00-dc264344edf0-logs" (OuterVolumeSpecName: "logs") pod "764ce995-4314-4c29-8b00-dc264344edf0" (UID: "764ce995-4314-4c29-8b00-dc264344edf0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.149408 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-scripts" (OuterVolumeSpecName: "scripts") pod "764ce995-4314-4c29-8b00-dc264344edf0" (UID: "764ce995-4314-4c29-8b00-dc264344edf0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.169582 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/764ce995-4314-4c29-8b00-dc264344edf0-kube-api-access-mvlqp" (OuterVolumeSpecName: "kube-api-access-mvlqp") pod "764ce995-4314-4c29-8b00-dc264344edf0" (UID: "764ce995-4314-4c29-8b00-dc264344edf0"). InnerVolumeSpecName "kube-api-access-mvlqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.183247 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-config-data" (OuterVolumeSpecName: "config-data") pod "764ce995-4314-4c29-8b00-dc264344edf0" (UID: "764ce995-4314-4c29-8b00-dc264344edf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.186496 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "764ce995-4314-4c29-8b00-dc264344edf0" (UID: "764ce995-4314-4c29-8b00-dc264344edf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.246419 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.246445 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvlqp\" (UniqueName: \"kubernetes.io/projected/764ce995-4314-4c29-8b00-dc264344edf0-kube-api-access-mvlqp\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.246455 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.246464 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/764ce995-4314-4c29-8b00-dc264344edf0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.246473 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/764ce995-4314-4c29-8b00-dc264344edf0-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.742124 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9499d9cb-dtnz8" event={"ID":"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae","Type":"ContainerStarted","Data":"cdd53a2da98b3590231182ee51d6fd17c12cef22ed6cd781d4b950ffaef044a6"} Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.742402 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.742413 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9499d9cb-dtnz8" event={"ID":"cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae","Type":"ContainerStarted","Data":"ceca4fc599f853821b20b47e1cef2f52decc4ea938875fae0fddfa7e68f0e3e2"} Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.745215 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-h76mn" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.745405 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-h76mn" event={"ID":"764ce995-4314-4c29-8b00-dc264344edf0","Type":"ContainerDied","Data":"d2970b231143622c6bdbc4f13c986140a5709797fcaef46757b2b3a1e4efe5ad"} Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.745450 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2970b231143622c6bdbc4f13c986140a5709797fcaef46757b2b3a1e4efe5ad" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.769212 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c9499d9cb-dtnz8" podStartSLOduration=1.769194843 podStartE2EDuration="1.769194843s" podCreationTimestamp="2025-11-24 19:33:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:09.767051014 +0000 UTC m=+1128.289557271" watchObservedRunningTime="2025-11-24 19:33:09.769194843 +0000 UTC m=+1128.291701100" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.809188 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6869cb8db6-rk7hs"] Nov 24 19:33:09 crc kubenswrapper[5035]: E1124 19:33:09.809563 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764ce995-4314-4c29-8b00-dc264344edf0" containerName="placement-db-sync" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.809576 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="764ce995-4314-4c29-8b00-dc264344edf0" containerName="placement-db-sync" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.809760 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="764ce995-4314-4c29-8b00-dc264344edf0" containerName="placement-db-sync" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.811003 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.814703 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.814932 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.815003 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sqtx8" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.815156 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.815353 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.837267 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6869cb8db6-rk7hs"] Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.856396 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-public-tls-certs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.856441 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-combined-ca-bundle\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.856483 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-internal-tls-certs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.856541 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-config-data\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.856568 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd5g9\" (UniqueName: \"kubernetes.io/projected/f9f396db-4624-4c15-91d5-47c22dbeb77c-kube-api-access-sd5g9\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.856586 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f396db-4624-4c15-91d5-47c22dbeb77c-logs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.856662 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-scripts\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.958651 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-public-tls-certs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.958689 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-combined-ca-bundle\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.958727 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-internal-tls-certs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.958762 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-config-data\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.958790 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd5g9\" (UniqueName: \"kubernetes.io/projected/f9f396db-4624-4c15-91d5-47c22dbeb77c-kube-api-access-sd5g9\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.958809 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f396db-4624-4c15-91d5-47c22dbeb77c-logs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.958863 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-scripts\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.961610 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9f396db-4624-4c15-91d5-47c22dbeb77c-logs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.966145 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-combined-ca-bundle\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.970700 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-public-tls-certs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.970900 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-scripts\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.970931 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-config-data\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.971208 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9f396db-4624-4c15-91d5-47c22dbeb77c-internal-tls-certs\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:09 crc kubenswrapper[5035]: I1124 19:33:09.980889 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd5g9\" (UniqueName: \"kubernetes.io/projected/f9f396db-4624-4c15-91d5-47c22dbeb77c-kube-api-access-sd5g9\") pod \"placement-6869cb8db6-rk7hs\" (UID: \"f9f396db-4624-4c15-91d5-47c22dbeb77c\") " pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.028575 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.059716 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-db-sync-config-data\") pod \"205bbe21-12fb-4cdf-bf2d-09008085f221\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.059808 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ckjt\" (UniqueName: \"kubernetes.io/projected/205bbe21-12fb-4cdf-bf2d-09008085f221-kube-api-access-6ckjt\") pod \"205bbe21-12fb-4cdf-bf2d-09008085f221\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.059854 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-combined-ca-bundle\") pod \"205bbe21-12fb-4cdf-bf2d-09008085f221\" (UID: \"205bbe21-12fb-4cdf-bf2d-09008085f221\") " Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.063661 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "205bbe21-12fb-4cdf-bf2d-09008085f221" (UID: "205bbe21-12fb-4cdf-bf2d-09008085f221"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.067204 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/205bbe21-12fb-4cdf-bf2d-09008085f221-kube-api-access-6ckjt" (OuterVolumeSpecName: "kube-api-access-6ckjt") pod "205bbe21-12fb-4cdf-bf2d-09008085f221" (UID: "205bbe21-12fb-4cdf-bf2d-09008085f221"). InnerVolumeSpecName "kube-api-access-6ckjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.113762 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "205bbe21-12fb-4cdf-bf2d-09008085f221" (UID: "205bbe21-12fb-4cdf-bf2d-09008085f221"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.127854 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.161946 5035 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.161972 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ckjt\" (UniqueName: \"kubernetes.io/projected/205bbe21-12fb-4cdf-bf2d-09008085f221-kube-api-access-6ckjt\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.161981 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205bbe21-12fb-4cdf-bf2d-09008085f221-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.680251 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6869cb8db6-rk7hs"] Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.760176 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6869cb8db6-rk7hs" event={"ID":"f9f396db-4624-4c15-91d5-47c22dbeb77c","Type":"ContainerStarted","Data":"da5815da8dd5831c1a93e4ad5116e93713932fc81a4fe65b08c7418780dc6ee3"} Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.773842 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xb7gt" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.773923 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xb7gt" event={"ID":"205bbe21-12fb-4cdf-bf2d-09008085f221","Type":"ContainerDied","Data":"1eb02672bfc9ed20132b6b3b0be8dfc5c2e03ed66cde46f6ae4c784372d6d179"} Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.773967 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1eb02672bfc9ed20132b6b3b0be8dfc5c2e03ed66cde46f6ae4c784372d6d179" Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.782726 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbbjc" event={"ID":"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f","Type":"ContainerStarted","Data":"37ef83088d4ee74d7be789192a3e95549cb61c54efe81b083affc749f3aff13c"} Nov 24 19:33:10 crc kubenswrapper[5035]: I1124 19:33:10.806369 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-nbbjc" podStartSLOduration=2.912178746 podStartE2EDuration="44.806352996s" podCreationTimestamp="2025-11-24 19:32:26 +0000 UTC" firstStartedPulling="2025-11-24 19:32:27.780778213 +0000 UTC m=+1086.303284470" lastFinishedPulling="2025-11-24 19:33:09.674952463 +0000 UTC m=+1128.197458720" observedRunningTime="2025-11-24 19:33:10.798850072 +0000 UTC m=+1129.321356319" watchObservedRunningTime="2025-11-24 19:33:10.806352996 +0000 UTC m=+1129.328859253" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.050338 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-9df689fb7-8fbwv"] Nov 24 19:33:11 crc kubenswrapper[5035]: E1124 19:33:11.051376 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205bbe21-12fb-4cdf-bf2d-09008085f221" containerName="barbican-db-sync" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.051394 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="205bbe21-12fb-4cdf-bf2d-09008085f221" containerName="barbican-db-sync" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.051683 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="205bbe21-12fb-4cdf-bf2d-09008085f221" containerName="barbican-db-sync" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.052961 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.055990 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.056187 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.059106 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-v6rv7" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.070933 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9df689fb7-8fbwv"] Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.096963 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-config-data\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.097012 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-combined-ca-bundle\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.097031 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/561f02ab-04d0-4100-8e25-66fd0df630f6-logs\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.097046 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-config-data-custom\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.097110 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qzxw\" (UniqueName: \"kubernetes.io/projected/561f02ab-04d0-4100-8e25-66fd0df630f6-kube-api-access-7qzxw\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.150161 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-bf699cb66-l5qkm"] Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.151588 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.154400 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.177341 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-bf699cb66-l5qkm"] Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201381 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-combined-ca-bundle\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201443 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qzxw\" (UniqueName: \"kubernetes.io/projected/561f02ab-04d0-4100-8e25-66fd0df630f6-kube-api-access-7qzxw\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201494 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cktx\" (UniqueName: \"kubernetes.io/projected/85943038-29af-4153-be73-f62fb83bf015-kube-api-access-5cktx\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201523 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-config-data-custom\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201549 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-config-data\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201565 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85943038-29af-4153-be73-f62fb83bf015-logs\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201582 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-config-data\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201609 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-combined-ca-bundle\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201626 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/561f02ab-04d0-4100-8e25-66fd0df630f6-logs\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.201644 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-config-data-custom\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.202466 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/561f02ab-04d0-4100-8e25-66fd0df630f6-logs\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.208139 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-combined-ca-bundle\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.212057 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-config-data\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.215908 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/561f02ab-04d0-4100-8e25-66fd0df630f6-config-data-custom\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.215992 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d548b9b8f-fjjwt"] Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.217480 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.224818 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qzxw\" (UniqueName: \"kubernetes.io/projected/561f02ab-04d0-4100-8e25-66fd0df630f6-kube-api-access-7qzxw\") pod \"barbican-worker-9df689fb7-8fbwv\" (UID: \"561f02ab-04d0-4100-8e25-66fd0df630f6\") " pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.253005 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d548b9b8f-fjjwt"] Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.277088 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-596cf4dbd-q478w"] Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.279201 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.281849 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.301780 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-596cf4dbd-q478w"] Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302738 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-config\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302789 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cktx\" (UniqueName: \"kubernetes.io/projected/85943038-29af-4153-be73-f62fb83bf015-kube-api-access-5cktx\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302827 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-config-data-custom\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302855 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85943038-29af-4153-be73-f62fb83bf015-logs\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302871 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-config-data\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302894 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h597\" (UniqueName: \"kubernetes.io/projected/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-kube-api-access-6h597\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302910 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-nb\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.302978 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-sb\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.303006 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-combined-ca-bundle\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.303031 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-dns-svc\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.306868 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-config-data-custom\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.307085 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85943038-29af-4153-be73-f62fb83bf015-logs\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.310577 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-config-data\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.314991 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85943038-29af-4153-be73-f62fb83bf015-combined-ca-bundle\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.320529 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cktx\" (UniqueName: \"kubernetes.io/projected/85943038-29af-4153-be73-f62fb83bf015-kube-api-access-5cktx\") pod \"barbican-keystone-listener-bf699cb66-l5qkm\" (UID: \"85943038-29af-4153-be73-f62fb83bf015\") " pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415070 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h597\" (UniqueName: \"kubernetes.io/projected/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-kube-api-access-6h597\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415123 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-nb\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415163 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9gzs\" (UniqueName: \"kubernetes.io/projected/5fcc53ad-89df-45ac-a3a3-37a7460633bc-kube-api-access-c9gzs\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415214 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415269 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fcc53ad-89df-45ac-a3a3-37a7460633bc-logs\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415315 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-sb\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415374 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-dns-svc\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415400 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data-custom\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415434 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-config\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.415501 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-combined-ca-bundle\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.416606 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-nb\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.417128 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-sb\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.417707 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-dns-svc\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.418205 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-config\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.436072 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-9df689fb7-8fbwv" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.444330 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h597\" (UniqueName: \"kubernetes.io/projected/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-kube-api-access-6h597\") pod \"dnsmasq-dns-7d548b9b8f-fjjwt\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.492724 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.516923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-combined-ca-bundle\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.517012 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9gzs\" (UniqueName: \"kubernetes.io/projected/5fcc53ad-89df-45ac-a3a3-37a7460633bc-kube-api-access-c9gzs\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.517043 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.517079 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fcc53ad-89df-45ac-a3a3-37a7460633bc-logs\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.517135 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data-custom\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.518685 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fcc53ad-89df-45ac-a3a3-37a7460633bc-logs\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.545446 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data-custom\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.546549 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.549793 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-combined-ca-bundle\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.550742 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9gzs\" (UniqueName: \"kubernetes.io/projected/5fcc53ad-89df-45ac-a3a3-37a7460633bc-kube-api-access-c9gzs\") pod \"barbican-api-596cf4dbd-q478w\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.691673 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.705702 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.810099 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6869cb8db6-rk7hs" event={"ID":"f9f396db-4624-4c15-91d5-47c22dbeb77c","Type":"ContainerStarted","Data":"be546e1d38a608d522fc054319a2323749df7898b1269a3633badf2e26137903"} Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.810344 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6869cb8db6-rk7hs" event={"ID":"f9f396db-4624-4c15-91d5-47c22dbeb77c","Type":"ContainerStarted","Data":"26c1332a751a14b1553aeaf740669b35cdb1851b4047e1079a1bebe7e272e95d"} Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.810454 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:11 crc kubenswrapper[5035]: I1124 19:33:11.857076 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6869cb8db6-rk7hs" podStartSLOduration=2.857053537 podStartE2EDuration="2.857053537s" podCreationTimestamp="2025-11-24 19:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:11.844060564 +0000 UTC m=+1130.366566821" watchObservedRunningTime="2025-11-24 19:33:11.857053537 +0000 UTC m=+1130.379559794" Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.033219 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-bf699cb66-l5qkm"] Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.158078 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-9df689fb7-8fbwv"] Nov 24 19:33:12 crc kubenswrapper[5035]: W1124 19:33:12.171143 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod561f02ab_04d0_4100_8e25_66fd0df630f6.slice/crio-f886f6bcd79a9e9732e60d6d2ccf7f4afdac0d01cca4d6e975755e490a5fbc57 WatchSource:0}: Error finding container f886f6bcd79a9e9732e60d6d2ccf7f4afdac0d01cca4d6e975755e490a5fbc57: Status 404 returned error can't find the container with id f886f6bcd79a9e9732e60d6d2ccf7f4afdac0d01cca4d6e975755e490a5fbc57 Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.319623 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d548b9b8f-fjjwt"] Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.349834 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-596cf4dbd-q478w"] Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.818604 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9df689fb7-8fbwv" event={"ID":"561f02ab-04d0-4100-8e25-66fd0df630f6","Type":"ContainerStarted","Data":"f886f6bcd79a9e9732e60d6d2ccf7f4afdac0d01cca4d6e975755e490a5fbc57"} Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.825124 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerID="001dcdf1a2f18240fd9019d3641e3e94fda9cf823872f71803a5d0a0881c216d" exitCode=0 Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.825194 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" event={"ID":"4d31ee82-2212-4b7f-a4d0-b58656ad2a59","Type":"ContainerDied","Data":"001dcdf1a2f18240fd9019d3641e3e94fda9cf823872f71803a5d0a0881c216d"} Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.825226 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" event={"ID":"4d31ee82-2212-4b7f-a4d0-b58656ad2a59","Type":"ContainerStarted","Data":"44da05ea0a7ec491a5a14490afe0056612916bc4bdba0fe30fcd66ef02568fa0"} Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.833934 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596cf4dbd-q478w" event={"ID":"5fcc53ad-89df-45ac-a3a3-37a7460633bc","Type":"ContainerStarted","Data":"33f8527618f935e5e4b7f1d0f2c77a90938d5f8dab5cf27356a9c5ebe845e5fe"} Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.833980 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596cf4dbd-q478w" event={"ID":"5fcc53ad-89df-45ac-a3a3-37a7460633bc","Type":"ContainerStarted","Data":"25c7c070eb352fd303fb29f17eb5e2b9f2f03b4689397ab16dbdcc06ff973d2c"} Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.833990 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596cf4dbd-q478w" event={"ID":"5fcc53ad-89df-45ac-a3a3-37a7460633bc","Type":"ContainerStarted","Data":"feec982c883c3055a02f7fb8085c969ef04358ec86d3585879a894a738637829"} Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.835167 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" event={"ID":"85943038-29af-4153-be73-f62fb83bf015","Type":"ContainerStarted","Data":"8b50e29cf1093a65b311f55184acbc66c1d17a412ae7b0102f2257b7b5ad8b1e"} Nov 24 19:33:12 crc kubenswrapper[5035]: I1124 19:33:12.835224 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.860431 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" event={"ID":"4d31ee82-2212-4b7f-a4d0-b58656ad2a59","Type":"ContainerStarted","Data":"eba3a9664768722b8adf41200183b47c1a23bac00d901d5beb96ef3fac1af23f"} Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.860794 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.860818 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.860832 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.862619 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-555d68d8cb-knjvm"] Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.866903 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.869868 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.870091 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.881527 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-555d68d8cb-knjvm"] Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.965804 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-596cf4dbd-q478w" podStartSLOduration=2.965784408 podStartE2EDuration="2.965784408s" podCreationTimestamp="2025-11-24 19:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:13.918405341 +0000 UTC m=+1132.440911598" watchObservedRunningTime="2025-11-24 19:33:13.965784408 +0000 UTC m=+1132.488290665" Nov 24 19:33:13 crc kubenswrapper[5035]: I1124 19:33:13.994558 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" podStartSLOduration=2.9945413690000002 podStartE2EDuration="2.994541369s" podCreationTimestamp="2025-11-24 19:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:13.989090241 +0000 UTC m=+1132.511596508" watchObservedRunningTime="2025-11-24 19:33:13.994541369 +0000 UTC m=+1132.517047626" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.009590 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-config-data-custom\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.009665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-public-tls-certs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.009712 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39154cf4-dfd5-4fb2-bd13-e9d15012da12-logs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.009742 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-config-data\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.009761 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr5lf\" (UniqueName: \"kubernetes.io/projected/39154cf4-dfd5-4fb2-bd13-e9d15012da12-kube-api-access-jr5lf\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.009839 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-internal-tls-certs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.009884 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-combined-ca-bundle\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.111099 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-config-data-custom\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.111160 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-public-tls-certs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.111190 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39154cf4-dfd5-4fb2-bd13-e9d15012da12-logs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.111212 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-config-data\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.111227 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr5lf\" (UniqueName: \"kubernetes.io/projected/39154cf4-dfd5-4fb2-bd13-e9d15012da12-kube-api-access-jr5lf\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.111272 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-internal-tls-certs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.111308 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-combined-ca-bundle\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.112337 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39154cf4-dfd5-4fb2-bd13-e9d15012da12-logs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.118045 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-combined-ca-bundle\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.118497 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-config-data-custom\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.119709 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-public-tls-certs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.119929 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-config-data\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.120728 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39154cf4-dfd5-4fb2-bd13-e9d15012da12-internal-tls-certs\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.132556 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr5lf\" (UniqueName: \"kubernetes.io/projected/39154cf4-dfd5-4fb2-bd13-e9d15012da12-kube-api-access-jr5lf\") pod \"barbican-api-555d68d8cb-knjvm\" (UID: \"39154cf4-dfd5-4fb2-bd13-e9d15012da12\") " pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:14 crc kubenswrapper[5035]: I1124 19:33:14.224673 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:15 crc kubenswrapper[5035]: I1124 19:33:15.234619 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:33:15 crc kubenswrapper[5035]: I1124 19:33:15.234679 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:33:16 crc kubenswrapper[5035]: I1124 19:33:16.724593 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6ff6c75f76-ttb9x" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 24 19:33:16 crc kubenswrapper[5035]: I1124 19:33:16.827988 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-677648b7f8-j8p2d" podUID="6c39b1fa-70b7-49f4-b318-0d08d5007dba" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Nov 24 19:33:17 crc kubenswrapper[5035]: I1124 19:33:17.892029 5035 generic.go:334] "Generic (PLEG): container finished" podID="0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" containerID="37ef83088d4ee74d7be789192a3e95549cb61c54efe81b083affc749f3aff13c" exitCode=0 Nov 24 19:33:17 crc kubenswrapper[5035]: I1124 19:33:17.892080 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbbjc" event={"ID":"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f","Type":"ContainerDied","Data":"37ef83088d4ee74d7be789192a3e95549cb61c54efe81b083affc749f3aff13c"} Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.339547 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.354861 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-596cf4dbd-q478w" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.417768 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-scripts\") pod \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.417893 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl5n6\" (UniqueName: \"kubernetes.io/projected/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-kube-api-access-dl5n6\") pod \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.417959 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-etc-machine-id\") pod \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.418170 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-combined-ca-bundle\") pod \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.418268 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-config-data\") pod \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.418356 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-db-sync-config-data\") pod \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\" (UID: \"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f\") " Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.422928 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" (UID: "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.425765 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-scripts" (OuterVolumeSpecName: "scripts") pod "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" (UID: "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.428397 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" (UID: "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.432857 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-kube-api-access-dl5n6" (OuterVolumeSpecName: "kube-api-access-dl5n6") pod "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" (UID: "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f"). InnerVolumeSpecName "kube-api-access-dl5n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.455277 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" (UID: "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.477716 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-config-data" (OuterVolumeSpecName: "config-data") pod "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" (UID: "0c9794ab-e4e3-4af0-a689-6ef761d1fe4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.520893 5035 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.520937 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.520947 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl5n6\" (UniqueName: \"kubernetes.io/projected/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-kube-api-access-dl5n6\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.520956 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.520964 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.520971 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:19 crc kubenswrapper[5035]: E1124 19:33:19.569537 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.676680 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-555d68d8cb-knjvm"] Nov 24 19:33:19 crc kubenswrapper[5035]: W1124 19:33:19.682003 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39154cf4_dfd5_4fb2_bd13_e9d15012da12.slice/crio-4dada4ab79de6a4dadd8c0e9ef665d9f4fe631686b2d9a0edc79f0c6902cb7e3 WatchSource:0}: Error finding container 4dada4ab79de6a4dadd8c0e9ef665d9f4fe631686b2d9a0edc79f0c6902cb7e3: Status 404 returned error can't find the container with id 4dada4ab79de6a4dadd8c0e9ef665d9f4fe631686b2d9a0edc79f0c6902cb7e3 Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.920438 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" event={"ID":"85943038-29af-4153-be73-f62fb83bf015","Type":"ContainerStarted","Data":"79dad1943332c5498f55572dd23539646b7a17cde8cccfb55e7ec4ec86958f89"} Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.925477 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbbjc" event={"ID":"0c9794ab-e4e3-4af0-a689-6ef761d1fe4f","Type":"ContainerDied","Data":"fd69585e1eb719c35886ba4568491c7b02dc4dc0456e059e4165bf77dde01e19"} Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.925517 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd69585e1eb719c35886ba4568491c7b02dc4dc0456e059e4165bf77dde01e19" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.925599 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbbjc" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.932145 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerStarted","Data":"8562e014c183a03d238bfd03743494b10344f00f09e1e21b2bdfb8d5f5ca58b3"} Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.932477 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="ceilometer-notification-agent" containerID="cri-o://8040e70ce1911937f52f82371500b42c5ac5ce93dfae37ee13dfafc26794c317" gracePeriod=30 Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.932611 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.933035 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="proxy-httpd" containerID="cri-o://8562e014c183a03d238bfd03743494b10344f00f09e1e21b2bdfb8d5f5ca58b3" gracePeriod=30 Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.933120 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="sg-core" containerID="cri-o://101366c589974592e10246af0ea4216583154d7bf940f6637349259623ae22c3" gracePeriod=30 Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.941790 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9df689fb7-8fbwv" event={"ID":"561f02ab-04d0-4100-8e25-66fd0df630f6","Type":"ContainerStarted","Data":"3d10f3786dc373a2a92239d529c46ea95a469d21602f54f1d880492a652da078"} Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.951254 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-555d68d8cb-knjvm" event={"ID":"39154cf4-dfd5-4fb2-bd13-e9d15012da12","Type":"ContainerStarted","Data":"f0216aefe6cc22314a48157cd64130bec893370d762c7c30d7d350ebc129a7ee"} Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.951312 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-555d68d8cb-knjvm" event={"ID":"39154cf4-dfd5-4fb2-bd13-e9d15012da12","Type":"ContainerStarted","Data":"4dada4ab79de6a4dadd8c0e9ef665d9f4fe631686b2d9a0edc79f0c6902cb7e3"} Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.953909 5035 generic.go:334] "Generic (PLEG): container finished" podID="c11a04cb-bd10-49c8-affa-5189af979040" containerID="4c6bb91f085a9a7f1b6c300c57b5ba3691a074db9a73773a414462dcd34bb968" exitCode=0 Nov 24 19:33:19 crc kubenswrapper[5035]: I1124 19:33:19.953965 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g7nvj" event={"ID":"c11a04cb-bd10-49c8-affa-5189af979040","Type":"ContainerDied","Data":"4c6bb91f085a9a7f1b6c300c57b5ba3691a074db9a73773a414462dcd34bb968"} Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.020181 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-9df689fb7-8fbwv" podStartSLOduration=2.8695863409999998 podStartE2EDuration="10.020159599s" podCreationTimestamp="2025-11-24 19:33:10 +0000 UTC" firstStartedPulling="2025-11-24 19:33:12.17423147 +0000 UTC m=+1130.696737727" lastFinishedPulling="2025-11-24 19:33:19.324804728 +0000 UTC m=+1137.847310985" observedRunningTime="2025-11-24 19:33:20.009441479 +0000 UTC m=+1138.531947746" watchObservedRunningTime="2025-11-24 19:33:20.020159599 +0000 UTC m=+1138.542665856" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.247726 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:20 crc kubenswrapper[5035]: E1124 19:33:20.248460 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" containerName="cinder-db-sync" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.248481 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" containerName="cinder-db-sync" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.248724 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" containerName="cinder-db-sync" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.249901 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.256181 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5mm55" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.256414 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.256536 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.256589 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.272057 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.334822 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdpn5\" (UniqueName: \"kubernetes.io/projected/31dbdd88-5a90-4340-a6d0-da10a996a36a-kube-api-access-gdpn5\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.334877 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.334921 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-scripts\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.334952 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.334999 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.335051 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31dbdd88-5a90-4340-a6d0-da10a996a36a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.352612 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d548b9b8f-fjjwt"] Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.353934 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" podUID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerName="dnsmasq-dns" containerID="cri-o://eba3a9664768722b8adf41200183b47c1a23bac00d901d5beb96ef3fac1af23f" gracePeriod=10 Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.355569 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.377644 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6ffb5bb857-5f8f6"] Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.381806 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.404096 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb5bb857-5f8f6"] Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437234 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437456 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-config\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437492 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31dbdd88-5a90-4340-a6d0-da10a996a36a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437521 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vst5w\" (UniqueName: \"kubernetes.io/projected/181d720d-60f1-48ce-bbcd-919888df478d-kube-api-access-vst5w\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437540 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-dns-svc\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437562 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdpn5\" (UniqueName: \"kubernetes.io/projected/31dbdd88-5a90-4340-a6d0-da10a996a36a-kube-api-access-gdpn5\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437584 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437600 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437639 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437658 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-scripts\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.437687 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.439445 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31dbdd88-5a90-4340-a6d0-da10a996a36a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.443486 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.443996 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.444306 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.451663 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-scripts\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.458601 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdpn5\" (UniqueName: \"kubernetes.io/projected/31dbdd88-5a90-4340-a6d0-da10a996a36a-kube-api-access-gdpn5\") pod \"cinder-scheduler-0\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.490524 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.493335 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.499336 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.502923 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541347 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-scripts\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541403 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541446 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541490 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541545 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data-custom\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541591 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-config\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541638 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vst5w\" (UniqueName: \"kubernetes.io/projected/181d720d-60f1-48ce-bbcd-919888df478d-kube-api-access-vst5w\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541664 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-dns-svc\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541685 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43dfee81-1c6d-462a-b0a1-385961dac4bc-logs\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541727 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541754 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43dfee81-1c6d-462a-b0a1-385961dac4bc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.541778 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqxt2\" (UniqueName: \"kubernetes.io/projected/43dfee81-1c6d-462a-b0a1-385961dac4bc-kube-api-access-pqxt2\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.542977 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.543554 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-config\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.556445 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.556589 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-dns-svc\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.565023 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vst5w\" (UniqueName: \"kubernetes.io/projected/181d720d-60f1-48ce-bbcd-919888df478d-kube-api-access-vst5w\") pod \"dnsmasq-dns-6ffb5bb857-5f8f6\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.581870 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.644720 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-scripts\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.644973 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.645065 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.645156 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data-custom\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.645304 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43dfee81-1c6d-462a-b0a1-385961dac4bc-logs\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.645387 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43dfee81-1c6d-462a-b0a1-385961dac4bc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.645447 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqxt2\" (UniqueName: \"kubernetes.io/projected/43dfee81-1c6d-462a-b0a1-385961dac4bc-kube-api-access-pqxt2\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.647819 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43dfee81-1c6d-462a-b0a1-385961dac4bc-logs\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.651049 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43dfee81-1c6d-462a-b0a1-385961dac4bc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.654406 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-scripts\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.658063 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data-custom\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.667048 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.674192 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.683913 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqxt2\" (UniqueName: \"kubernetes.io/projected/43dfee81-1c6d-462a-b0a1-385961dac4bc-kube-api-access-pqxt2\") pod \"cinder-api-0\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " pod="openstack/cinder-api-0" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.756754 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:20 crc kubenswrapper[5035]: I1124 19:33:20.816700 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.002464 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" event={"ID":"85943038-29af-4153-be73-f62fb83bf015","Type":"ContainerStarted","Data":"7dd23b6f4df883d7ff0160811a6191c87dac8ea71c968f0990e3cb0c95558bdf"} Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.055228 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-bf699cb66-l5qkm" podStartSLOduration=2.8679335310000003 podStartE2EDuration="10.055213715s" podCreationTimestamp="2025-11-24 19:33:11 +0000 UTC" firstStartedPulling="2025-11-24 19:33:12.047666653 +0000 UTC m=+1130.570172910" lastFinishedPulling="2025-11-24 19:33:19.234946837 +0000 UTC m=+1137.757453094" observedRunningTime="2025-11-24 19:33:21.042993864 +0000 UTC m=+1139.565500121" watchObservedRunningTime="2025-11-24 19:33:21.055213715 +0000 UTC m=+1139.577719972" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.078537 5035 generic.go:334] "Generic (PLEG): container finished" podID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerID="8562e014c183a03d238bfd03743494b10344f00f09e1e21b2bdfb8d5f5ca58b3" exitCode=0 Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.078593 5035 generic.go:334] "Generic (PLEG): container finished" podID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerID="101366c589974592e10246af0ea4216583154d7bf940f6637349259623ae22c3" exitCode=2 Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.078675 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerDied","Data":"8562e014c183a03d238bfd03743494b10344f00f09e1e21b2bdfb8d5f5ca58b3"} Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.078717 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerDied","Data":"101366c589974592e10246af0ea4216583154d7bf940f6637349259623ae22c3"} Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.121586 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-9df689fb7-8fbwv" event={"ID":"561f02ab-04d0-4100-8e25-66fd0df630f6","Type":"ContainerStarted","Data":"7f1f7bd562692a10679d473956077d3b922779b398c70ea5cd2a1081604446d8"} Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.145679 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.161684 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerID="eba3a9664768722b8adf41200183b47c1a23bac00d901d5beb96ef3fac1af23f" exitCode=0 Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.161771 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" event={"ID":"4d31ee82-2212-4b7f-a4d0-b58656ad2a59","Type":"ContainerDied","Data":"eba3a9664768722b8adf41200183b47c1a23bac00d901d5beb96ef3fac1af23f"} Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.161816 5035 scope.go:117] "RemoveContainer" containerID="eba3a9664768722b8adf41200183b47c1a23bac00d901d5beb96ef3fac1af23f" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.191542 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-555d68d8cb-knjvm" event={"ID":"39154cf4-dfd5-4fb2-bd13-e9d15012da12","Type":"ContainerStarted","Data":"0d40cc3e3398b0510ff3347e0dd4fecd4121ce1521029c79b463215b925a420b"} Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.191574 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.191595 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.222441 5035 scope.go:117] "RemoveContainer" containerID="001dcdf1a2f18240fd9019d3641e3e94fda9cf823872f71803a5d0a0881c216d" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.260879 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-555d68d8cb-knjvm" podStartSLOduration=8.260862889 podStartE2EDuration="8.260862889s" podCreationTimestamp="2025-11-24 19:33:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:21.248961577 +0000 UTC m=+1139.771467834" watchObservedRunningTime="2025-11-24 19:33:21.260862889 +0000 UTC m=+1139.783369146" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.261505 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h597\" (UniqueName: \"kubernetes.io/projected/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-kube-api-access-6h597\") pod \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.261561 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-sb\") pod \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.261600 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-config\") pod \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.261687 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-dns-svc\") pod \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.261848 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-nb\") pod \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\" (UID: \"4d31ee82-2212-4b7f-a4d0-b58656ad2a59\") " Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.270463 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-kube-api-access-6h597" (OuterVolumeSpecName: "kube-api-access-6h597") pod "4d31ee82-2212-4b7f-a4d0-b58656ad2a59" (UID: "4d31ee82-2212-4b7f-a4d0-b58656ad2a59"). InnerVolumeSpecName "kube-api-access-6h597". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.346183 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4d31ee82-2212-4b7f-a4d0-b58656ad2a59" (UID: "4d31ee82-2212-4b7f-a4d0-b58656ad2a59"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.361030 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d31ee82-2212-4b7f-a4d0-b58656ad2a59" (UID: "4d31ee82-2212-4b7f-a4d0-b58656ad2a59"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.390628 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h597\" (UniqueName: \"kubernetes.io/projected/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-kube-api-access-6h597\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.390869 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.390879 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.403475 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.432114 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-config" (OuterVolumeSpecName: "config") pod "4d31ee82-2212-4b7f-a4d0-b58656ad2a59" (UID: "4d31ee82-2212-4b7f-a4d0-b58656ad2a59"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.440647 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4d31ee82-2212-4b7f-a4d0-b58656ad2a59" (UID: "4d31ee82-2212-4b7f-a4d0-b58656ad2a59"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.492926 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.492977 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d31ee82-2212-4b7f-a4d0-b58656ad2a59-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.791223 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb5bb857-5f8f6"] Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.822737 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:21 crc kubenswrapper[5035]: I1124 19:33:21.990799 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.104984 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqksl\" (UniqueName: \"kubernetes.io/projected/c11a04cb-bd10-49c8-affa-5189af979040-kube-api-access-tqksl\") pod \"c11a04cb-bd10-49c8-affa-5189af979040\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.105050 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-config\") pod \"c11a04cb-bd10-49c8-affa-5189af979040\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.105131 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-combined-ca-bundle\") pod \"c11a04cb-bd10-49c8-affa-5189af979040\" (UID: \"c11a04cb-bd10-49c8-affa-5189af979040\") " Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.123485 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c11a04cb-bd10-49c8-affa-5189af979040-kube-api-access-tqksl" (OuterVolumeSpecName: "kube-api-access-tqksl") pod "c11a04cb-bd10-49c8-affa-5189af979040" (UID: "c11a04cb-bd10-49c8-affa-5189af979040"). InnerVolumeSpecName "kube-api-access-tqksl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.133691 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-config" (OuterVolumeSpecName: "config") pod "c11a04cb-bd10-49c8-affa-5189af979040" (UID: "c11a04cb-bd10-49c8-affa-5189af979040"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.138269 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c11a04cb-bd10-49c8-affa-5189af979040" (UID: "c11a04cb-bd10-49c8-affa-5189af979040"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.233129 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.233159 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqksl\" (UniqueName: \"kubernetes.io/projected/c11a04cb-bd10-49c8-affa-5189af979040-kube-api-access-tqksl\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.233171 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c11a04cb-bd10-49c8-affa-5189af979040-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.281473 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"43dfee81-1c6d-462a-b0a1-385961dac4bc","Type":"ContainerStarted","Data":"14c464c7b9290307762884f234cef829a79d05d66ae8665ab0afad49bf120fd7"} Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.295048 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb5bb857-5f8f6"] Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.313757 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" event={"ID":"181d720d-60f1-48ce-bbcd-919888df478d","Type":"ContainerStarted","Data":"391eaa0b751d76a7b21f3a6ffbffe5b8e55c8496cdff6a7763bd5c4bf187459a"} Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.338483 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-g7nvj" event={"ID":"c11a04cb-bd10-49c8-affa-5189af979040","Type":"ContainerDied","Data":"cae340075a47e28d8ed89b9c6b4a4c708894e1f900d3f92d54a523155e84e9c9"} Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.338520 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cae340075a47e28d8ed89b9c6b4a4c708894e1f900d3f92d54a523155e84e9c9" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.338577 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-g7nvj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.393793 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" event={"ID":"4d31ee82-2212-4b7f-a4d0-b58656ad2a59","Type":"ContainerDied","Data":"44da05ea0a7ec491a5a14490afe0056612916bc4bdba0fe30fcd66ef02568fa0"} Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.393884 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d548b9b8f-fjjwt" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.415552 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"31dbdd88-5a90-4340-a6d0-da10a996a36a","Type":"ContainerStarted","Data":"41c4d1f2b94bde8846d0486a4cbcde99a664f1823554f8cc53b19dc4df2d3024"} Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.429113 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bc89f58d7-sc6h8"] Nov 24 19:33:22 crc kubenswrapper[5035]: E1124 19:33:22.429471 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerName="init" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.429482 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerName="init" Nov 24 19:33:22 crc kubenswrapper[5035]: E1124 19:33:22.429498 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerName="dnsmasq-dns" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.429504 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerName="dnsmasq-dns" Nov 24 19:33:22 crc kubenswrapper[5035]: E1124 19:33:22.429513 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11a04cb-bd10-49c8-affa-5189af979040" containerName="neutron-db-sync" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.429518 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11a04cb-bd10-49c8-affa-5189af979040" containerName="neutron-db-sync" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.429673 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" containerName="dnsmasq-dns" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.438354 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c11a04cb-bd10-49c8-affa-5189af979040" containerName="neutron-db-sync" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.439328 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.483358 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bc89f58d7-sc6h8"] Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.541483 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxh4m\" (UniqueName: \"kubernetes.io/projected/c6cf7f09-7707-4a27-a61e-1d49897c2e63-kube-api-access-vxh4m\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.541837 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.541866 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-config\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.541962 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-dns-svc\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.542001 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.588081 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6f878cb596-jgfmj"] Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.589619 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.598878 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.599151 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.599376 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qbgkv" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.599682 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.634359 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f878cb596-jgfmj"] Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.644561 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxh4m\" (UniqueName: \"kubernetes.io/projected/c6cf7f09-7707-4a27-a61e-1d49897c2e63-kube-api-access-vxh4m\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.644902 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.644963 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-config\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.645045 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-combined-ca-bundle\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.645100 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-dns-svc\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.645161 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.645229 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.645264 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-httpd-config\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.645350 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-config\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.645387 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4p7g\" (UniqueName: \"kubernetes.io/projected/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-kube-api-access-w4p7g\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.762487 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-combined-ca-bundle\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.762824 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.762907 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-httpd-config\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.762990 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-config\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.763074 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4p7g\" (UniqueName: \"kubernetes.io/projected/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-kube-api-access-w4p7g\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.784119 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d548b9b8f-fjjwt"] Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.806318 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d548b9b8f-fjjwt"] Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.988997 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.989252 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-dns-svc\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.989852 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-config\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:22 crc kubenswrapper[5035]: I1124 19:33:22.990388 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.000538 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxh4m\" (UniqueName: \"kubernetes.io/projected/c6cf7f09-7707-4a27-a61e-1d49897c2e63-kube-api-access-vxh4m\") pod \"dnsmasq-dns-7bc89f58d7-sc6h8\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.015102 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-httpd-config\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.015197 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4p7g\" (UniqueName: \"kubernetes.io/projected/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-kube-api-access-w4p7g\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.016012 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.016241 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-combined-ca-bundle\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.016365 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-config\") pod \"neutron-6f878cb596-jgfmj\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.201610 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.202459 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.218021 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.468797 5035 generic.go:334] "Generic (PLEG): container finished" podID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerID="8040e70ce1911937f52f82371500b42c5ac5ce93dfae37ee13dfafc26794c317" exitCode=0 Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.468875 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerDied","Data":"8040e70ce1911937f52f82371500b42c5ac5ce93dfae37ee13dfafc26794c317"} Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.472433 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"43dfee81-1c6d-462a-b0a1-385961dac4bc","Type":"ContainerStarted","Data":"ccf4a0e0b42fedc31fb53bc6d40a67778a0a2669fbfdb25768a73a834342f95d"} Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.482470 5035 generic.go:334] "Generic (PLEG): container finished" podID="181d720d-60f1-48ce-bbcd-919888df478d" containerID="c03576b36e8b6afe20901bf2870ac5357f8d63d9275d31e9f33dd0aba63242f2" exitCode=0 Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.482622 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" event={"ID":"181d720d-60f1-48ce-bbcd-919888df478d","Type":"ContainerDied","Data":"c03576b36e8b6afe20901bf2870ac5357f8d63d9275d31e9f33dd0aba63242f2"} Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.587512 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.682885 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7md9\" (UniqueName: \"kubernetes.io/projected/f9739bc1-965c-4f67-9b14-b96e176d72c1-kube-api-access-k7md9\") pod \"f9739bc1-965c-4f67-9b14-b96e176d72c1\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.683513 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-log-httpd\") pod \"f9739bc1-965c-4f67-9b14-b96e176d72c1\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.683548 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-sg-core-conf-yaml\") pod \"f9739bc1-965c-4f67-9b14-b96e176d72c1\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.683583 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-scripts\") pod \"f9739bc1-965c-4f67-9b14-b96e176d72c1\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.683638 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-config-data\") pod \"f9739bc1-965c-4f67-9b14-b96e176d72c1\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.683655 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-combined-ca-bundle\") pod \"f9739bc1-965c-4f67-9b14-b96e176d72c1\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.683690 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-run-httpd\") pod \"f9739bc1-965c-4f67-9b14-b96e176d72c1\" (UID: \"f9739bc1-965c-4f67-9b14-b96e176d72c1\") " Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.684358 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f9739bc1-965c-4f67-9b14-b96e176d72c1" (UID: "f9739bc1-965c-4f67-9b14-b96e176d72c1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.697504 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f9739bc1-965c-4f67-9b14-b96e176d72c1" (UID: "f9739bc1-965c-4f67-9b14-b96e176d72c1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.706910 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-scripts" (OuterVolumeSpecName: "scripts") pod "f9739bc1-965c-4f67-9b14-b96e176d72c1" (UID: "f9739bc1-965c-4f67-9b14-b96e176d72c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.713276 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9739bc1-965c-4f67-9b14-b96e176d72c1-kube-api-access-k7md9" (OuterVolumeSpecName: "kube-api-access-k7md9") pod "f9739bc1-965c-4f67-9b14-b96e176d72c1" (UID: "f9739bc1-965c-4f67-9b14-b96e176d72c1"). InnerVolumeSpecName "kube-api-access-k7md9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.785624 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.785876 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.785993 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9739bc1-965c-4f67-9b14-b96e176d72c1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.786067 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7md9\" (UniqueName: \"kubernetes.io/projected/f9739bc1-965c-4f67-9b14-b96e176d72c1-kube-api-access-k7md9\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.871180 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f9739bc1-965c-4f67-9b14-b96e176d72c1" (UID: "f9739bc1-965c-4f67-9b14-b96e176d72c1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.879787 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9739bc1-965c-4f67-9b14-b96e176d72c1" (UID: "f9739bc1-965c-4f67-9b14-b96e176d72c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.887598 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.887641 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:23 crc kubenswrapper[5035]: I1124 19:33:23.887971 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-config-data" (OuterVolumeSpecName: "config-data") pod "f9739bc1-965c-4f67-9b14-b96e176d72c1" (UID: "f9739bc1-965c-4f67-9b14-b96e176d72c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.005859 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9739bc1-965c-4f67-9b14-b96e176d72c1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.070567 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bc89f58d7-sc6h8"] Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.116951 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.210010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vst5w\" (UniqueName: \"kubernetes.io/projected/181d720d-60f1-48ce-bbcd-919888df478d-kube-api-access-vst5w\") pod \"181d720d-60f1-48ce-bbcd-919888df478d\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.210070 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-nb\") pod \"181d720d-60f1-48ce-bbcd-919888df478d\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.210092 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-dns-svc\") pod \"181d720d-60f1-48ce-bbcd-919888df478d\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.210158 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-sb\") pod \"181d720d-60f1-48ce-bbcd-919888df478d\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.210196 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-config\") pod \"181d720d-60f1-48ce-bbcd-919888df478d\" (UID: \"181d720d-60f1-48ce-bbcd-919888df478d\") " Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.230494 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181d720d-60f1-48ce-bbcd-919888df478d-kube-api-access-vst5w" (OuterVolumeSpecName: "kube-api-access-vst5w") pod "181d720d-60f1-48ce-bbcd-919888df478d" (UID: "181d720d-60f1-48ce-bbcd-919888df478d"). InnerVolumeSpecName "kube-api-access-vst5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.237138 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d31ee82-2212-4b7f-a4d0-b58656ad2a59" path="/var/lib/kubelet/pods/4d31ee82-2212-4b7f-a4d0-b58656ad2a59/volumes" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.243821 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-config" (OuterVolumeSpecName: "config") pod "181d720d-60f1-48ce-bbcd-919888df478d" (UID: "181d720d-60f1-48ce-bbcd-919888df478d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.275622 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "181d720d-60f1-48ce-bbcd-919888df478d" (UID: "181d720d-60f1-48ce-bbcd-919888df478d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.288130 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "181d720d-60f1-48ce-bbcd-919888df478d" (UID: "181d720d-60f1-48ce-bbcd-919888df478d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.296370 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "181d720d-60f1-48ce-bbcd-919888df478d" (UID: "181d720d-60f1-48ce-bbcd-919888df478d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.313205 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vst5w\" (UniqueName: \"kubernetes.io/projected/181d720d-60f1-48ce-bbcd-919888df478d-kube-api-access-vst5w\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.313236 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.313246 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.313255 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.313265 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/181d720d-60f1-48ce-bbcd-919888df478d-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.496407 5035 generic.go:334] "Generic (PLEG): container finished" podID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerID="67fe3d6d4e1fa113da4420884d9c72448baa4c96f66f8d353ad7d8790bce4c3a" exitCode=0 Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.496482 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" event={"ID":"c6cf7f09-7707-4a27-a61e-1d49897c2e63","Type":"ContainerDied","Data":"67fe3d6d4e1fa113da4420884d9c72448baa4c96f66f8d353ad7d8790bce4c3a"} Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.496509 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" event={"ID":"c6cf7f09-7707-4a27-a61e-1d49897c2e63","Type":"ContainerStarted","Data":"e98a6aef6f8ead4b2f901f27bcaa6a22ae99eb7219ce29ab781ffdc569acfe2a"} Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.512798 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.513072 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9739bc1-965c-4f67-9b14-b96e176d72c1","Type":"ContainerDied","Data":"fb54ffc05444118e954d391982ce8bc1d1485cd98d3b5f37dd69b0255068a973"} Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.513206 5035 scope.go:117] "RemoveContainer" containerID="8562e014c183a03d238bfd03743494b10344f00f09e1e21b2bdfb8d5f5ca58b3" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.515731 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" event={"ID":"181d720d-60f1-48ce-bbcd-919888df478d","Type":"ContainerDied","Data":"391eaa0b751d76a7b21f3a6ffbffe5b8e55c8496cdff6a7763bd5c4bf187459a"} Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.516252 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb5bb857-5f8f6" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.548171 5035 scope.go:117] "RemoveContainer" containerID="101366c589974592e10246af0ea4216583154d7bf940f6637349259623ae22c3" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.577818 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.586012 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.620362 5035 scope.go:117] "RemoveContainer" containerID="8040e70ce1911937f52f82371500b42c5ac5ce93dfae37ee13dfafc26794c317" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763002 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:33:24 crc kubenswrapper[5035]: E1124 19:33:24.763607 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="sg-core" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763624 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="sg-core" Nov 24 19:33:24 crc kubenswrapper[5035]: E1124 19:33:24.763634 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="ceilometer-notification-agent" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763640 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="ceilometer-notification-agent" Nov 24 19:33:24 crc kubenswrapper[5035]: E1124 19:33:24.763665 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="proxy-httpd" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763671 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="proxy-httpd" Nov 24 19:33:24 crc kubenswrapper[5035]: E1124 19:33:24.763686 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181d720d-60f1-48ce-bbcd-919888df478d" containerName="init" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763692 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="181d720d-60f1-48ce-bbcd-919888df478d" containerName="init" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763836 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="181d720d-60f1-48ce-bbcd-919888df478d" containerName="init" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763860 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="ceilometer-notification-agent" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763869 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="sg-core" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.763879 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" containerName="proxy-httpd" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.765362 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.774817 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.775116 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.792664 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb5bb857-5f8f6"] Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.798555 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.814556 5035 scope.go:117] "RemoveContainer" containerID="c03576b36e8b6afe20901bf2870ac5357f8d63d9275d31e9f33dd0aba63242f2" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.827642 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.827913 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-scripts\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.828056 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckzsv\" (UniqueName: \"kubernetes.io/projected/0b594afb-a6dd-42d4-9430-91e0da323e72-kube-api-access-ckzsv\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.828238 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-config-data\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.828438 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-run-httpd\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.828557 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.828770 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-log-httpd\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.829749 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6ffb5bb857-5f8f6"] Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.839328 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f878cb596-jgfmj"] Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.938216 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckzsv\" (UniqueName: \"kubernetes.io/projected/0b594afb-a6dd-42d4-9430-91e0da323e72-kube-api-access-ckzsv\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.938264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-config-data\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.949737 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-run-httpd\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.950043 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.950075 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-log-httpd\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.950323 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.950374 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-scripts\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.952640 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-log-httpd\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.953111 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-run-httpd\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.954763 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.957991 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-scripts\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.963934 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.964249 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.964919 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckzsv\" (UniqueName: \"kubernetes.io/projected/0b594afb-a6dd-42d4-9430-91e0da323e72-kube-api-access-ckzsv\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.965312 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:24 crc kubenswrapper[5035]: I1124 19:33:24.965630 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-config-data\") pod \"ceilometer-0\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " pod="openstack/ceilometer-0" Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.113734 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.529313 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" event={"ID":"c6cf7f09-7707-4a27-a61e-1d49897c2e63","Type":"ContainerStarted","Data":"53794eb118a5ab08840029397a1db927b614e05877a303e0e0a52c339c579b93"} Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.530535 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.537928 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"31dbdd88-5a90-4340-a6d0-da10a996a36a","Type":"ContainerStarted","Data":"795ace031cc268ff382f7ae4084e3347250121bb0ebf13e76bfdb18b0a32481a"} Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.565635 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"43dfee81-1c6d-462a-b0a1-385961dac4bc","Type":"ContainerStarted","Data":"1eca89228143957c73fbcee91b6f27e773b574c10ae66556eb5605a76ce35f07"} Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.565694 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.565698 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api-log" containerID="cri-o://ccf4a0e0b42fedc31fb53bc6d40a67778a0a2669fbfdb25768a73a834342f95d" gracePeriod=30 Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.565729 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api" containerID="cri-o://1eca89228143957c73fbcee91b6f27e773b574c10ae66556eb5605a76ce35f07" gracePeriod=30 Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.569651 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f878cb596-jgfmj" event={"ID":"4bb5614e-a370-47a0-8cf8-2f4fec8f8537","Type":"ContainerStarted","Data":"799fe7e3633104ea6e51df226bddf64e6732760c5381fcb03a067fec8ebf0a6a"} Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.589069 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" podStartSLOduration=3.589052959 podStartE2EDuration="3.589052959s" podCreationTimestamp="2025-11-24 19:33:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:25.548623541 +0000 UTC m=+1144.071129798" watchObservedRunningTime="2025-11-24 19:33:25.589052959 +0000 UTC m=+1144.111559216" Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.756703 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.75668193 podStartE2EDuration="5.75668193s" podCreationTimestamp="2025-11-24 19:33:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:25.599897152 +0000 UTC m=+1144.122403419" watchObservedRunningTime="2025-11-24 19:33:25.75668193 +0000 UTC m=+1144.279188187" Nov 24 19:33:25 crc kubenswrapper[5035]: I1124 19:33:25.767657 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:33:25 crc kubenswrapper[5035]: W1124 19:33:25.774590 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b594afb_a6dd_42d4_9430_91e0da323e72.slice/crio-7b8921a6b50b4e44f034b26d6791df8cf8f1a30932596ae64f3ccc5bf673024c WatchSource:0}: Error finding container 7b8921a6b50b4e44f034b26d6791df8cf8f1a30932596ae64f3ccc5bf673024c: Status 404 returned error can't find the container with id 7b8921a6b50b4e44f034b26d6791df8cf8f1a30932596ae64f3ccc5bf673024c Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.213611 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="181d720d-60f1-48ce-bbcd-919888df478d" path="/var/lib/kubelet/pods/181d720d-60f1-48ce-bbcd-919888df478d/volumes" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.214388 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9739bc1-965c-4f67-9b14-b96e176d72c1" path="/var/lib/kubelet/pods/f9739bc1-965c-4f67-9b14-b96e176d72c1/volumes" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.315807 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.581406 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f878cb596-jgfmj" event={"ID":"4bb5614e-a370-47a0-8cf8-2f4fec8f8537","Type":"ContainerStarted","Data":"5903b708b5f8b8569216d549b4ed3b3a0261ee7bfab5325abb5b93eedd2751e8"} Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.581447 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f878cb596-jgfmj" event={"ID":"4bb5614e-a370-47a0-8cf8-2f4fec8f8537","Type":"ContainerStarted","Data":"6238cf869c5881652c52e47d931c34742b8e65e50c69f4130ea134e344a5d74a"} Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.582447 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.586870 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerStarted","Data":"7b8921a6b50b4e44f034b26d6791df8cf8f1a30932596ae64f3ccc5bf673024c"} Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.592493 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"31dbdd88-5a90-4340-a6d0-da10a996a36a","Type":"ContainerStarted","Data":"1c5865aebd892b5e2a1f226e4f1b20486505783cfedf1aec7ee517b18972dc81"} Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.608674 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6f878cb596-jgfmj" podStartSLOduration=4.608658905 podStartE2EDuration="4.608658905s" podCreationTimestamp="2025-11-24 19:33:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:26.604769529 +0000 UTC m=+1145.127275786" watchObservedRunningTime="2025-11-24 19:33:26.608658905 +0000 UTC m=+1145.131165162" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.612687 5035 generic.go:334] "Generic (PLEG): container finished" podID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerID="1eca89228143957c73fbcee91b6f27e773b574c10ae66556eb5605a76ce35f07" exitCode=0 Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.612726 5035 generic.go:334] "Generic (PLEG): container finished" podID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerID="ccf4a0e0b42fedc31fb53bc6d40a67778a0a2669fbfdb25768a73a834342f95d" exitCode=143 Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.613102 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"43dfee81-1c6d-462a-b0a1-385961dac4bc","Type":"ContainerDied","Data":"1eca89228143957c73fbcee91b6f27e773b574c10ae66556eb5605a76ce35f07"} Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.613172 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"43dfee81-1c6d-462a-b0a1-385961dac4bc","Type":"ContainerDied","Data":"ccf4a0e0b42fedc31fb53bc6d40a67778a0a2669fbfdb25768a73a834342f95d"} Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.628411 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.553078948 podStartE2EDuration="6.628389451s" podCreationTimestamp="2025-11-24 19:33:20 +0000 UTC" firstStartedPulling="2025-11-24 19:33:21.392964157 +0000 UTC m=+1139.915470414" lastFinishedPulling="2025-11-24 19:33:23.46827466 +0000 UTC m=+1141.990780917" observedRunningTime="2025-11-24 19:33:26.622198462 +0000 UTC m=+1145.144704719" watchObservedRunningTime="2025-11-24 19:33:26.628389451 +0000 UTC m=+1145.150895708" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.803596 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6b7c585949-9h76s"] Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.805823 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.812360 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b7c585949-9h76s"] Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.815428 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.815712 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.913628 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7qnn\" (UniqueName: \"kubernetes.io/projected/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-kube-api-access-z7qnn\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.913671 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-combined-ca-bundle\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.913723 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-config\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.913749 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-public-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.913779 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-internal-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.913813 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-httpd-config\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:26 crc kubenswrapper[5035]: I1124 19:33:26.913839 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-ovndb-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.019394 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-internal-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.019771 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-httpd-config\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.019815 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-ovndb-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.019877 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7qnn\" (UniqueName: \"kubernetes.io/projected/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-kube-api-access-z7qnn\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.019914 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-combined-ca-bundle\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.019988 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-config\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.020030 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-public-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.027352 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-ovndb-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.027673 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-combined-ca-bundle\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.028248 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-httpd-config\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.028407 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-internal-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.029838 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-config\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.037212 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-public-tls-certs\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.056985 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7qnn\" (UniqueName: \"kubernetes.io/projected/6886d0f1-8b8b-4580-8145-f5bb671cdf1f-kube-api-access-z7qnn\") pod \"neutron-6b7c585949-9h76s\" (UID: \"6886d0f1-8b8b-4580-8145-f5bb671cdf1f\") " pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.107175 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.225806 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43dfee81-1c6d-462a-b0a1-385961dac4bc-logs\") pod \"43dfee81-1c6d-462a-b0a1-385961dac4bc\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.225850 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43dfee81-1c6d-462a-b0a1-385961dac4bc-etc-machine-id\") pod \"43dfee81-1c6d-462a-b0a1-385961dac4bc\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.225929 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data-custom\") pod \"43dfee81-1c6d-462a-b0a1-385961dac4bc\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.225975 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data\") pod \"43dfee81-1c6d-462a-b0a1-385961dac4bc\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.226013 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqxt2\" (UniqueName: \"kubernetes.io/projected/43dfee81-1c6d-462a-b0a1-385961dac4bc-kube-api-access-pqxt2\") pod \"43dfee81-1c6d-462a-b0a1-385961dac4bc\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.226070 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-scripts\") pod \"43dfee81-1c6d-462a-b0a1-385961dac4bc\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.226086 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-combined-ca-bundle\") pod \"43dfee81-1c6d-462a-b0a1-385961dac4bc\" (UID: \"43dfee81-1c6d-462a-b0a1-385961dac4bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.239404 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43dfee81-1c6d-462a-b0a1-385961dac4bc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "43dfee81-1c6d-462a-b0a1-385961dac4bc" (UID: "43dfee81-1c6d-462a-b0a1-385961dac4bc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.239574 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43dfee81-1c6d-462a-b0a1-385961dac4bc-logs" (OuterVolumeSpecName: "logs") pod "43dfee81-1c6d-462a-b0a1-385961dac4bc" (UID: "43dfee81-1c6d-462a-b0a1-385961dac4bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.244226 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "43dfee81-1c6d-462a-b0a1-385961dac4bc" (UID: "43dfee81-1c6d-462a-b0a1-385961dac4bc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.256596 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43dfee81-1c6d-462a-b0a1-385961dac4bc-kube-api-access-pqxt2" (OuterVolumeSpecName: "kube-api-access-pqxt2") pod "43dfee81-1c6d-462a-b0a1-385961dac4bc" (UID: "43dfee81-1c6d-462a-b0a1-385961dac4bc"). InnerVolumeSpecName "kube-api-access-pqxt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.259430 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-scripts" (OuterVolumeSpecName: "scripts") pod "43dfee81-1c6d-462a-b0a1-385961dac4bc" (UID: "43dfee81-1c6d-462a-b0a1-385961dac4bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.281762 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.287425 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43dfee81-1c6d-462a-b0a1-385961dac4bc" (UID: "43dfee81-1c6d-462a-b0a1-385961dac4bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.322357 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.329326 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43dfee81-1c6d-462a-b0a1-385961dac4bc-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.329346 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43dfee81-1c6d-462a-b0a1-385961dac4bc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.329356 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.329365 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqxt2\" (UniqueName: \"kubernetes.io/projected/43dfee81-1c6d-462a-b0a1-385961dac4bc-kube-api-access-pqxt2\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.329374 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.329382 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.362525 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data" (OuterVolumeSpecName: "config-data") pod "43dfee81-1c6d-462a-b0a1-385961dac4bc" (UID: "43dfee81-1c6d-462a-b0a1-385961dac4bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.430340 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-config-data\") pod \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.430413 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7c1bf22-059d-45fc-9db2-1bcb791949bc-horizon-secret-key\") pod \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.430538 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c1bf22-059d-45fc-9db2-1bcb791949bc-logs\") pod \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.430629 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-scripts\") pod \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.430667 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tstkp\" (UniqueName: \"kubernetes.io/projected/e7c1bf22-059d-45fc-9db2-1bcb791949bc-kube-api-access-tstkp\") pod \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\" (UID: \"e7c1bf22-059d-45fc-9db2-1bcb791949bc\") " Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.431005 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43dfee81-1c6d-462a-b0a1-385961dac4bc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.430992 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7c1bf22-059d-45fc-9db2-1bcb791949bc-logs" (OuterVolumeSpecName: "logs") pod "e7c1bf22-059d-45fc-9db2-1bcb791949bc" (UID: "e7c1bf22-059d-45fc-9db2-1bcb791949bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.438449 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7c1bf22-059d-45fc-9db2-1bcb791949bc-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e7c1bf22-059d-45fc-9db2-1bcb791949bc" (UID: "e7c1bf22-059d-45fc-9db2-1bcb791949bc"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.438500 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7c1bf22-059d-45fc-9db2-1bcb791949bc-kube-api-access-tstkp" (OuterVolumeSpecName: "kube-api-access-tstkp") pod "e7c1bf22-059d-45fc-9db2-1bcb791949bc" (UID: "e7c1bf22-059d-45fc-9db2-1bcb791949bc"). InnerVolumeSpecName "kube-api-access-tstkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.458518 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-config-data" (OuterVolumeSpecName: "config-data") pod "e7c1bf22-059d-45fc-9db2-1bcb791949bc" (UID: "e7c1bf22-059d-45fc-9db2-1bcb791949bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.468128 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-scripts" (OuterVolumeSpecName: "scripts") pod "e7c1bf22-059d-45fc-9db2-1bcb791949bc" (UID: "e7c1bf22-059d-45fc-9db2-1bcb791949bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.534572 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c1bf22-059d-45fc-9db2-1bcb791949bc-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.534605 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.534614 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tstkp\" (UniqueName: \"kubernetes.io/projected/e7c1bf22-059d-45fc-9db2-1bcb791949bc-kube-api-access-tstkp\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.534624 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7c1bf22-059d-45fc-9db2-1bcb791949bc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.534633 5035 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e7c1bf22-059d-45fc-9db2-1bcb791949bc-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.632858 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerStarted","Data":"a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76"} Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.634568 5035 generic.go:334] "Generic (PLEG): container finished" podID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerID="50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a" exitCode=137 Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.634605 5035 generic.go:334] "Generic (PLEG): container finished" podID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerID="0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7" exitCode=137 Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.634631 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97bc6dcf7-zpd2w" event={"ID":"e7c1bf22-059d-45fc-9db2-1bcb791949bc","Type":"ContainerDied","Data":"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a"} Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.634673 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97bc6dcf7-zpd2w" event={"ID":"e7c1bf22-059d-45fc-9db2-1bcb791949bc","Type":"ContainerDied","Data":"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7"} Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.634685 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-97bc6dcf7-zpd2w" event={"ID":"e7c1bf22-059d-45fc-9db2-1bcb791949bc","Type":"ContainerDied","Data":"937d4283fa54c6ec32690a40cfb2c0c050e6b3b06e81eb901f17b582c9a907c2"} Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.634685 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-97bc6dcf7-zpd2w" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.635708 5035 scope.go:117] "RemoveContainer" containerID="50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.638071 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.640630 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"43dfee81-1c6d-462a-b0a1-385961dac4bc","Type":"ContainerDied","Data":"14c464c7b9290307762884f234cef829a79d05d66ae8665ab0afad49bf120fd7"} Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.894943 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-97bc6dcf7-zpd2w"] Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.930259 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-97bc6dcf7-zpd2w"] Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.942432 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.954353 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.973360 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:27 crc kubenswrapper[5035]: E1124 19:33:27.973792 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.973814 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api" Nov 24 19:33:27 crc kubenswrapper[5035]: E1124 19:33:27.973827 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api-log" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.973834 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api-log" Nov 24 19:33:27 crc kubenswrapper[5035]: E1124 19:33:27.973856 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon-log" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.973862 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon-log" Nov 24 19:33:27 crc kubenswrapper[5035]: E1124 19:33:27.973874 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.973880 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.974019 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.974040 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" containerName="cinder-api-log" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.974051 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon-log" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.974068 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" containerName="horizon" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.975106 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.982797 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.983015 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.983171 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.984017 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:27 crc kubenswrapper[5035]: I1124 19:33:27.988694 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b7c585949-9h76s"] Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.045389 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-config-data-custom\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046080 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046150 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-scripts\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046177 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37be962-03c8-4e0a-b329-12d6765d5c77-logs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046243 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a37be962-03c8-4e0a-b329-12d6765d5c77-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046265 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046349 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046464 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-config-data\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.046530 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976zb\" (UniqueName: \"kubernetes.io/projected/a37be962-03c8-4e0a-b329-12d6765d5c77-kube-api-access-976zb\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.057598 5035 scope.go:117] "RemoveContainer" containerID="0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150269 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150595 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-config-data\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150648 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976zb\" (UniqueName: \"kubernetes.io/projected/a37be962-03c8-4e0a-b329-12d6765d5c77-kube-api-access-976zb\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150675 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-config-data-custom\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150702 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150741 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-scripts\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150758 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37be962-03c8-4e0a-b329-12d6765d5c77-logs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150796 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a37be962-03c8-4e0a-b329-12d6765d5c77-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.150810 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.151109 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a37be962-03c8-4e0a-b329-12d6765d5c77-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.151306 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37be962-03c8-4e0a-b329-12d6765d5c77-logs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.155782 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-config-data\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.155802 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.155901 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.156601 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-config-data-custom\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.156901 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.158121 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a37be962-03c8-4e0a-b329-12d6765d5c77-scripts\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.168673 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976zb\" (UniqueName: \"kubernetes.io/projected/a37be962-03c8-4e0a-b329-12d6765d5c77-kube-api-access-976zb\") pod \"cinder-api-0\" (UID: \"a37be962-03c8-4e0a-b329-12d6765d5c77\") " pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.186174 5035 scope.go:117] "RemoveContainer" containerID="50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a" Nov 24 19:33:28 crc kubenswrapper[5035]: E1124 19:33:28.197096 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a\": container with ID starting with 50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a not found: ID does not exist" containerID="50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.197149 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a"} err="failed to get container status \"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a\": rpc error: code = NotFound desc = could not find container \"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a\": container with ID starting with 50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a not found: ID does not exist" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.197175 5035 scope.go:117] "RemoveContainer" containerID="0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7" Nov 24 19:33:28 crc kubenswrapper[5035]: E1124 19:33:28.197592 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7\": container with ID starting with 0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7 not found: ID does not exist" containerID="0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.197622 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7"} err="failed to get container status \"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7\": rpc error: code = NotFound desc = could not find container \"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7\": container with ID starting with 0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7 not found: ID does not exist" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.197635 5035 scope.go:117] "RemoveContainer" containerID="50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.197980 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a"} err="failed to get container status \"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a\": rpc error: code = NotFound desc = could not find container \"50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a\": container with ID starting with 50945711c16f7aba5a7ea94dda00b46f1036eaf2bbcfd93715d5e95c6d0b637a not found: ID does not exist" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.198003 5035 scope.go:117] "RemoveContainer" containerID="0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.198229 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7"} err="failed to get container status \"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7\": rpc error: code = NotFound desc = could not find container \"0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7\": container with ID starting with 0b03f1fde246dcd4bd63b6b7553332d9dc88ade82426d3fd37045e83c7c2acb7 not found: ID does not exist" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.198256 5035 scope.go:117] "RemoveContainer" containerID="1eca89228143957c73fbcee91b6f27e773b574c10ae66556eb5605a76ce35f07" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.229620 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43dfee81-1c6d-462a-b0a1-385961dac4bc" path="/var/lib/kubelet/pods/43dfee81-1c6d-462a-b0a1-385961dac4bc/volumes" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.230946 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7c1bf22-059d-45fc-9db2-1bcb791949bc" path="/var/lib/kubelet/pods/e7c1bf22-059d-45fc-9db2-1bcb791949bc/volumes" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.242471 5035 scope.go:117] "RemoveContainer" containerID="ccf4a0e0b42fedc31fb53bc6d40a67778a0a2669fbfdb25768a73a834342f95d" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.351364 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.671031 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b7c585949-9h76s" event={"ID":"6886d0f1-8b8b-4580-8145-f5bb671cdf1f","Type":"ContainerStarted","Data":"bd6ec84e59ff68ced8cbfbfa1f7f992058beb55bbaa117234311320e322e1111"} Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.671250 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b7c585949-9h76s" event={"ID":"6886d0f1-8b8b-4580-8145-f5bb671cdf1f","Type":"ContainerStarted","Data":"2f8c7678cddc5263b9ca2105d2078013c2703795fc5e930bc44a0e951b64f415"} Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.674437 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerStarted","Data":"a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8"} Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.938494 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 19:33:28 crc kubenswrapper[5035]: I1124 19:33:28.957262 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-555d68d8cb-knjvm" Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.026520 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-596cf4dbd-q478w"] Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.026767 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-596cf4dbd-q478w" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api-log" containerID="cri-o://25c7c070eb352fd303fb29f17eb5e2b9f2f03b4689397ab16dbdcc06ff973d2c" gracePeriod=30 Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.030731 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-596cf4dbd-q478w" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api" containerID="cri-o://33f8527618f935e5e4b7f1d0f2c77a90938d5f8dab5cf27356a9c5ebe845e5fe" gracePeriod=30 Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.709741 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.712765 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b7c585949-9h76s" event={"ID":"6886d0f1-8b8b-4580-8145-f5bb671cdf1f","Type":"ContainerStarted","Data":"4a418936c9444d1e8e814635829c7cab50456550d6c3cbd41331848ecda4b927"} Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.713249 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.716344 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerStarted","Data":"b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76"} Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.725775 5035 generic.go:334] "Generic (PLEG): container finished" podID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerID="25c7c070eb352fd303fb29f17eb5e2b9f2f03b4689397ab16dbdcc06ff973d2c" exitCode=143 Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.725824 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596cf4dbd-q478w" event={"ID":"5fcc53ad-89df-45ac-a3a3-37a7460633bc","Type":"ContainerDied","Data":"25c7c070eb352fd303fb29f17eb5e2b9f2f03b4689397ab16dbdcc06ff973d2c"} Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.726860 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a37be962-03c8-4e0a-b329-12d6765d5c77","Type":"ContainerStarted","Data":"7dc31dd5680711ed460439331feb7d5dbaaf4cf93a373ca364ea30a47327ad59"} Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.896265 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:33:29 crc kubenswrapper[5035]: I1124 19:33:29.919711 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6b7c585949-9h76s" podStartSLOduration=3.919691924 podStartE2EDuration="3.919691924s" podCreationTimestamp="2025-11-24 19:33:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:29.754522779 +0000 UTC m=+1148.277029036" watchObservedRunningTime="2025-11-24 19:33:29.919691924 +0000 UTC m=+1148.442198181" Nov 24 19:33:30 crc kubenswrapper[5035]: I1124 19:33:30.583552 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 19:33:30 crc kubenswrapper[5035]: I1124 19:33:30.737417 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a37be962-03c8-4e0a-b329-12d6765d5c77","Type":"ContainerStarted","Data":"ba4e1b20b96cf4dfca72a31868db1da5ed25077fd80bfd4b5e6c272ab8dd1048"} Nov 24 19:33:30 crc kubenswrapper[5035]: I1124 19:33:30.737457 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a37be962-03c8-4e0a-b329-12d6765d5c77","Type":"ContainerStarted","Data":"162d3befd3808b66de3976cf57a46364ee469045cf64b633da4882cb7479be63"} Nov 24 19:33:30 crc kubenswrapper[5035]: I1124 19:33:30.762282 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.762261503 podStartE2EDuration="3.762261503s" podCreationTimestamp="2025-11-24 19:33:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:30.755340035 +0000 UTC m=+1149.277846292" watchObservedRunningTime="2025-11-24 19:33:30.762261503 +0000 UTC m=+1149.284767760" Nov 24 19:33:30 crc kubenswrapper[5035]: I1124 19:33:30.898315 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 19:33:30 crc kubenswrapper[5035]: I1124 19:33:30.938514 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:31 crc kubenswrapper[5035]: I1124 19:33:31.750435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerStarted","Data":"62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d"} Nov 24 19:33:31 crc kubenswrapper[5035]: I1124 19:33:31.750671 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="cinder-scheduler" containerID="cri-o://795ace031cc268ff382f7ae4084e3347250121bb0ebf13e76bfdb18b0a32481a" gracePeriod=30 Nov 24 19:33:31 crc kubenswrapper[5035]: I1124 19:33:31.750826 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="probe" containerID="cri-o://1c5865aebd892b5e2a1f226e4f1b20486505783cfedf1aec7ee517b18972dc81" gracePeriod=30 Nov 24 19:33:31 crc kubenswrapper[5035]: I1124 19:33:31.750932 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 19:33:31 crc kubenswrapper[5035]: I1124 19:33:31.751556 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.029995 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.054369 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.3279425959999998 podStartE2EDuration="8.054348518s" podCreationTimestamp="2025-11-24 19:33:24 +0000 UTC" firstStartedPulling="2025-11-24 19:33:25.776485838 +0000 UTC m=+1144.298992095" lastFinishedPulling="2025-11-24 19:33:30.50289176 +0000 UTC m=+1149.025398017" observedRunningTime="2025-11-24 19:33:31.780462152 +0000 UTC m=+1150.302968399" watchObservedRunningTime="2025-11-24 19:33:32.054348518 +0000 UTC m=+1150.576854775" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.283179 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-677648b7f8-j8p2d" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.324226 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-596cf4dbd-q478w" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.150:9311/healthcheck\": read tcp 10.217.0.2:44894->10.217.0.150:9311: read: connection reset by peer" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.324571 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-596cf4dbd-q478w" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.150:9311/healthcheck\": read tcp 10.217.0.2:44880->10.217.0.150:9311: read: connection reset by peer" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.335596 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ff6c75f76-ttb9x"] Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.770831 5035 generic.go:334] "Generic (PLEG): container finished" podID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerID="33f8527618f935e5e4b7f1d0f2c77a90938d5f8dab5cf27356a9c5ebe845e5fe" exitCode=0 Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.771060 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596cf4dbd-q478w" event={"ID":"5fcc53ad-89df-45ac-a3a3-37a7460633bc","Type":"ContainerDied","Data":"33f8527618f935e5e4b7f1d0f2c77a90938d5f8dab5cf27356a9c5ebe845e5fe"} Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.772619 5035 generic.go:334] "Generic (PLEG): container finished" podID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerID="1c5865aebd892b5e2a1f226e4f1b20486505783cfedf1aec7ee517b18972dc81" exitCode=0 Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.772639 5035 generic.go:334] "Generic (PLEG): container finished" podID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerID="795ace031cc268ff382f7ae4084e3347250121bb0ebf13e76bfdb18b0a32481a" exitCode=0 Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.772792 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6ff6c75f76-ttb9x" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon-log" containerID="cri-o://df984fe11df0c4b1f22de9b25a3845a96fb7f38e7c96e121c2987ad75c805c5c" gracePeriod=30 Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.773002 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"31dbdd88-5a90-4340-a6d0-da10a996a36a","Type":"ContainerDied","Data":"1c5865aebd892b5e2a1f226e4f1b20486505783cfedf1aec7ee517b18972dc81"} Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.773023 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"31dbdd88-5a90-4340-a6d0-da10a996a36a","Type":"ContainerDied","Data":"795ace031cc268ff382f7ae4084e3347250121bb0ebf13e76bfdb18b0a32481a"} Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.777253 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6ff6c75f76-ttb9x" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" containerID="cri-o://b273b803cecfebabd01ad70881af43d61fe52dfee628924d6091e2d362e9ca4c" gracePeriod=30 Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.875071 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.920062 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.952071 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdpn5\" (UniqueName: \"kubernetes.io/projected/31dbdd88-5a90-4340-a6d0-da10a996a36a-kube-api-access-gdpn5\") pod \"31dbdd88-5a90-4340-a6d0-da10a996a36a\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.952155 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-scripts\") pod \"31dbdd88-5a90-4340-a6d0-da10a996a36a\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.952284 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31dbdd88-5a90-4340-a6d0-da10a996a36a-etc-machine-id\") pod \"31dbdd88-5a90-4340-a6d0-da10a996a36a\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.952343 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data\") pod \"31dbdd88-5a90-4340-a6d0-da10a996a36a\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.952475 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data-custom\") pod \"31dbdd88-5a90-4340-a6d0-da10a996a36a\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.952535 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-combined-ca-bundle\") pod \"31dbdd88-5a90-4340-a6d0-da10a996a36a\" (UID: \"31dbdd88-5a90-4340-a6d0-da10a996a36a\") " Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.953055 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31dbdd88-5a90-4340-a6d0-da10a996a36a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "31dbdd88-5a90-4340-a6d0-da10a996a36a" (UID: "31dbdd88-5a90-4340-a6d0-da10a996a36a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.961651 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31dbdd88-5a90-4340-a6d0-da10a996a36a-kube-api-access-gdpn5" (OuterVolumeSpecName: "kube-api-access-gdpn5") pod "31dbdd88-5a90-4340-a6d0-da10a996a36a" (UID: "31dbdd88-5a90-4340-a6d0-da10a996a36a"). InnerVolumeSpecName "kube-api-access-gdpn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.967409 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-scripts" (OuterVolumeSpecName: "scripts") pod "31dbdd88-5a90-4340-a6d0-da10a996a36a" (UID: "31dbdd88-5a90-4340-a6d0-da10a996a36a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:32 crc kubenswrapper[5035]: I1124 19:33:32.979086 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "31dbdd88-5a90-4340-a6d0-da10a996a36a" (UID: "31dbdd88-5a90-4340-a6d0-da10a996a36a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.016866 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31dbdd88-5a90-4340-a6d0-da10a996a36a" (UID: "31dbdd88-5a90-4340-a6d0-da10a996a36a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.055909 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-combined-ca-bundle\") pod \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.056064 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9gzs\" (UniqueName: \"kubernetes.io/projected/5fcc53ad-89df-45ac-a3a3-37a7460633bc-kube-api-access-c9gzs\") pod \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.056214 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fcc53ad-89df-45ac-a3a3-37a7460633bc-logs\") pod \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.056251 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data-custom\") pod \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.056360 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data\") pod \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\" (UID: \"5fcc53ad-89df-45ac-a3a3-37a7460633bc\") " Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.056975 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fcc53ad-89df-45ac-a3a3-37a7460633bc-logs" (OuterVolumeSpecName: "logs") pod "5fcc53ad-89df-45ac-a3a3-37a7460633bc" (UID: "5fcc53ad-89df-45ac-a3a3-37a7460633bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.057468 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fcc53ad-89df-45ac-a3a3-37a7460633bc-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.057502 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31dbdd88-5a90-4340-a6d0-da10a996a36a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.057518 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.057531 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.057542 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdpn5\" (UniqueName: \"kubernetes.io/projected/31dbdd88-5a90-4340-a6d0-da10a996a36a-kube-api-access-gdpn5\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.057554 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.062421 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5fcc53ad-89df-45ac-a3a3-37a7460633bc" (UID: "5fcc53ad-89df-45ac-a3a3-37a7460633bc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.062800 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data" (OuterVolumeSpecName: "config-data") pod "31dbdd88-5a90-4340-a6d0-da10a996a36a" (UID: "31dbdd88-5a90-4340-a6d0-da10a996a36a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.064521 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fcc53ad-89df-45ac-a3a3-37a7460633bc-kube-api-access-c9gzs" (OuterVolumeSpecName: "kube-api-access-c9gzs") pod "5fcc53ad-89df-45ac-a3a3-37a7460633bc" (UID: "5fcc53ad-89df-45ac-a3a3-37a7460633bc"). InnerVolumeSpecName "kube-api-access-c9gzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.089344 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fcc53ad-89df-45ac-a3a3-37a7460633bc" (UID: "5fcc53ad-89df-45ac-a3a3-37a7460633bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.110467 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data" (OuterVolumeSpecName: "config-data") pod "5fcc53ad-89df-45ac-a3a3-37a7460633bc" (UID: "5fcc53ad-89df-45ac-a3a3-37a7460633bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.162560 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.162604 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9gzs\" (UniqueName: \"kubernetes.io/projected/5fcc53ad-89df-45ac-a3a3-37a7460633bc-kube-api-access-c9gzs\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.162619 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.162634 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31dbdd88-5a90-4340-a6d0-da10a996a36a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.162645 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcc53ad-89df-45ac-a3a3-37a7460633bc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.204478 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.284526 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f8f5cc67-rtdk2"] Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.286564 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" podUID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerName="dnsmasq-dns" containerID="cri-o://c6764f8e05c3e71625b817583ec3c97fbccff709414175e646f4845b99af1ce6" gracePeriod=10 Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.781222 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-596cf4dbd-q478w" event={"ID":"5fcc53ad-89df-45ac-a3a3-37a7460633bc","Type":"ContainerDied","Data":"feec982c883c3055a02f7fb8085c969ef04358ec86d3585879a894a738637829"} Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.781248 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-596cf4dbd-q478w" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.781267 5035 scope.go:117] "RemoveContainer" containerID="33f8527618f935e5e4b7f1d0f2c77a90938d5f8dab5cf27356a9c5ebe845e5fe" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.783522 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"31dbdd88-5a90-4340-a6d0-da10a996a36a","Type":"ContainerDied","Data":"41c4d1f2b94bde8846d0486a4cbcde99a664f1823554f8cc53b19dc4df2d3024"} Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.783548 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.793930 5035 generic.go:334] "Generic (PLEG): container finished" podID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerID="c6764f8e05c3e71625b817583ec3c97fbccff709414175e646f4845b99af1ce6" exitCode=0 Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.793976 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" event={"ID":"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48","Type":"ContainerDied","Data":"c6764f8e05c3e71625b817583ec3c97fbccff709414175e646f4845b99af1ce6"} Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.808349 5035 scope.go:117] "RemoveContainer" containerID="25c7c070eb352fd303fb29f17eb5e2b9f2f03b4689397ab16dbdcc06ff973d2c" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.819894 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-596cf4dbd-q478w"] Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.829462 5035 scope.go:117] "RemoveContainer" containerID="1c5865aebd892b5e2a1f226e4f1b20486505783cfedf1aec7ee517b18972dc81" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.838383 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-596cf4dbd-q478w"] Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.850040 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.857258 5035 scope.go:117] "RemoveContainer" containerID="795ace031cc268ff382f7ae4084e3347250121bb0ebf13e76bfdb18b0a32481a" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.858039 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.874394 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:33 crc kubenswrapper[5035]: E1124 19:33:33.874724 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.874736 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api" Nov 24 19:33:33 crc kubenswrapper[5035]: E1124 19:33:33.874810 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api-log" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.874818 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api-log" Nov 24 19:33:33 crc kubenswrapper[5035]: E1124 19:33:33.874828 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="cinder-scheduler" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.874835 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="cinder-scheduler" Nov 24 19:33:33 crc kubenswrapper[5035]: E1124 19:33:33.874847 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="probe" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.874888 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="probe" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.875036 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.875049 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="cinder-scheduler" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.875059 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" containerName="probe" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.875073 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" containerName="barbican-api-log" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.876016 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.880579 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.890568 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.978836 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.978907 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgqz\" (UniqueName: \"kubernetes.io/projected/a1d160b0-7b77-4f8c-a918-170911d5ca2a-kube-api-access-plgqz\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.979039 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d160b0-7b77-4f8c-a918-170911d5ca2a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.979081 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-scripts\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.979109 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-config-data\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:33 crc kubenswrapper[5035]: I1124 19:33:33.979138 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.080857 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.081213 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgqz\" (UniqueName: \"kubernetes.io/projected/a1d160b0-7b77-4f8c-a918-170911d5ca2a-kube-api-access-plgqz\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.081630 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d160b0-7b77-4f8c-a918-170911d5ca2a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.081678 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-scripts\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.081701 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-config-data\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.081729 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.083298 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d160b0-7b77-4f8c-a918-170911d5ca2a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.086238 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.086418 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.097886 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-config-data\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.113917 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d160b0-7b77-4f8c-a918-170911d5ca2a-scripts\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.116077 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgqz\" (UniqueName: \"kubernetes.io/projected/a1d160b0-7b77-4f8c-a918-170911d5ca2a-kube-api-access-plgqz\") pod \"cinder-scheduler-0\" (UID: \"a1d160b0-7b77-4f8c-a918-170911d5ca2a\") " pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.202393 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.230548 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31dbdd88-5a90-4340-a6d0-da10a996a36a" path="/var/lib/kubelet/pods/31dbdd88-5a90-4340-a6d0-da10a996a36a/volumes" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.231175 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fcc53ad-89df-45ac-a3a3-37a7460633bc" path="/var/lib/kubelet/pods/5fcc53ad-89df-45ac-a3a3-37a7460633bc/volumes" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.437724 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.486800 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-config\") pod \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.486862 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mgk2\" (UniqueName: \"kubernetes.io/projected/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-kube-api-access-6mgk2\") pod \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.486935 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-nb\") pod \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.487011 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-sb\") pod \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.487175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-dns-svc\") pod \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\" (UID: \"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48\") " Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.515511 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-kube-api-access-6mgk2" (OuterVolumeSpecName: "kube-api-access-6mgk2") pod "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" (UID: "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48"). InnerVolumeSpecName "kube-api-access-6mgk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.567002 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" (UID: "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.567500 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" (UID: "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.582248 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" (UID: "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.589116 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mgk2\" (UniqueName: \"kubernetes.io/projected/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-kube-api-access-6mgk2\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.589155 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.589164 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.589173 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.597451 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-config" (OuterVolumeSpecName: "config") pod "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" (UID: "1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.691345 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.791445 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.814860 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" event={"ID":"1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48","Type":"ContainerDied","Data":"8b331ffbbc33898aabf40c6b2d1443db8ece40e8d576a189fa2f7f4c46bb3fd3"} Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.814909 5035 scope.go:117] "RemoveContainer" containerID="c6764f8e05c3e71625b817583ec3c97fbccff709414175e646f4845b99af1ce6" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.814914 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f8f5cc67-rtdk2" Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.820824 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1d160b0-7b77-4f8c-a918-170911d5ca2a","Type":"ContainerStarted","Data":"6bbdee5bfb3700dd6aa602ee2e7ad094cc20b1a8c48718f92347a47d60442b95"} Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.856098 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f8f5cc67-rtdk2"] Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.862601 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f8f5cc67-rtdk2"] Nov 24 19:33:34 crc kubenswrapper[5035]: I1124 19:33:34.863120 5035 scope.go:117] "RemoveContainer" containerID="5b70d778c68167893f7706062c30d8e3ec0ca777807f27307e521175b2315545" Nov 24 19:33:35 crc kubenswrapper[5035]: I1124 19:33:35.834410 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1d160b0-7b77-4f8c-a918-170911d5ca2a","Type":"ContainerStarted","Data":"f76c2495f1389d533f19d22857ecd3b9beceba0ae843aa0d29b057ae1d56aafa"} Nov 24 19:33:36 crc kubenswrapper[5035]: I1124 19:33:36.212826 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" path="/var/lib/kubelet/pods/1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48/volumes" Nov 24 19:33:36 crc kubenswrapper[5035]: I1124 19:33:36.723992 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6ff6c75f76-ttb9x" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 24 19:33:36 crc kubenswrapper[5035]: I1124 19:33:36.846913 5035 generic.go:334] "Generic (PLEG): container finished" podID="e67a302e-93be-4341-a7e1-a52183398050" containerID="b273b803cecfebabd01ad70881af43d61fe52dfee628924d6091e2d362e9ca4c" exitCode=0 Nov 24 19:33:36 crc kubenswrapper[5035]: I1124 19:33:36.846991 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff6c75f76-ttb9x" event={"ID":"e67a302e-93be-4341-a7e1-a52183398050","Type":"ContainerDied","Data":"b273b803cecfebabd01ad70881af43d61fe52dfee628924d6091e2d362e9ca4c"} Nov 24 19:33:36 crc kubenswrapper[5035]: I1124 19:33:36.848994 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1d160b0-7b77-4f8c-a918-170911d5ca2a","Type":"ContainerStarted","Data":"ab897a880217333bad855f22d3d30d27f8a9540629718dbc9289edfe25a77cb9"} Nov 24 19:33:36 crc kubenswrapper[5035]: I1124 19:33:36.881929 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.881911307 podStartE2EDuration="3.881911307s" podCreationTimestamp="2025-11-24 19:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:36.880557741 +0000 UTC m=+1155.403063998" watchObservedRunningTime="2025-11-24 19:33:36.881911307 +0000 UTC m=+1155.404417574" Nov 24 19:33:39 crc kubenswrapper[5035]: I1124 19:33:39.203644 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 19:33:40 crc kubenswrapper[5035]: I1124 19:33:40.210662 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 19:33:40 crc kubenswrapper[5035]: I1124 19:33:40.283665 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-c9499d9cb-dtnz8" Nov 24 19:33:41 crc kubenswrapper[5035]: I1124 19:33:41.752811 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:41 crc kubenswrapper[5035]: I1124 19:33:41.771437 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6869cb8db6-rk7hs" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.956750 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 19:33:43 crc kubenswrapper[5035]: E1124 19:33:43.957466 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerName="dnsmasq-dns" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.957481 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerName="dnsmasq-dns" Nov 24 19:33:43 crc kubenswrapper[5035]: E1124 19:33:43.957519 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerName="init" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.957526 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerName="init" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.957708 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2bb2c4-3cf8-4385-a5a3-e8576b6a0e48" containerName="dnsmasq-dns" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.958428 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.967607 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.967608 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xq2t4" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.967678 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 19:33:43 crc kubenswrapper[5035]: I1124 19:33:43.989327 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.090151 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1128a14c-1553-433f-b64b-676f3cec11f9-openstack-config\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.090226 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1128a14c-1553-433f-b64b-676f3cec11f9-openstack-config-secret\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.090325 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwgsf\" (UniqueName: \"kubernetes.io/projected/1128a14c-1553-433f-b64b-676f3cec11f9-kube-api-access-kwgsf\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.090348 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1128a14c-1553-433f-b64b-676f3cec11f9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.191891 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwgsf\" (UniqueName: \"kubernetes.io/projected/1128a14c-1553-433f-b64b-676f3cec11f9-kube-api-access-kwgsf\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.191962 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1128a14c-1553-433f-b64b-676f3cec11f9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.192014 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1128a14c-1553-433f-b64b-676f3cec11f9-openstack-config\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.192083 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1128a14c-1553-433f-b64b-676f3cec11f9-openstack-config-secret\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.194012 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1128a14c-1553-433f-b64b-676f3cec11f9-openstack-config\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.208333 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1128a14c-1553-433f-b64b-676f3cec11f9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.208608 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1128a14c-1553-433f-b64b-676f3cec11f9-openstack-config-secret\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.215954 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwgsf\" (UniqueName: \"kubernetes.io/projected/1128a14c-1553-433f-b64b-676f3cec11f9-kube-api-access-kwgsf\") pod \"openstackclient\" (UID: \"1128a14c-1553-433f-b64b-676f3cec11f9\") " pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.287632 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.459610 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.807512 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 19:33:44 crc kubenswrapper[5035]: I1124 19:33:44.923273 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1128a14c-1553-433f-b64b-676f3cec11f9","Type":"ContainerStarted","Data":"f578db6eeb8415c87962b5d08a0e77c7dda62d0f779fafa0befebe3ec518b2fd"} Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.234755 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.234841 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.234911 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.235724 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2a1e9879f1117f6209b6479ae153039678ddcd07bf4e0e5400eeea2f3c473b0"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.235793 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://d2a1e9879f1117f6209b6479ae153039678ddcd07bf4e0e5400eeea2f3c473b0" gracePeriod=600 Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.934642 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="d2a1e9879f1117f6209b6479ae153039678ddcd07bf4e0e5400eeea2f3c473b0" exitCode=0 Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.934810 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"d2a1e9879f1117f6209b6479ae153039678ddcd07bf4e0e5400eeea2f3c473b0"} Nov 24 19:33:45 crc kubenswrapper[5035]: I1124 19:33:45.935010 5035 scope.go:117] "RemoveContainer" containerID="7f8c76c75a096d1e632ca8dbc062a2c86c922da2cd856db7b41a1460c00d789d" Nov 24 19:33:46 crc kubenswrapper[5035]: I1124 19:33:46.723274 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6ff6c75f76-ttb9x" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 24 19:33:46 crc kubenswrapper[5035]: I1124 19:33:46.947175 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"5e56edadec21ff9db609cb1eb46281d662ed2c26ae0bd15afd911d69fb3661e1"} Nov 24 19:33:52 crc kubenswrapper[5035]: I1124 19:33:52.715394 5035 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod4d31ee82-2212-4b7f-a4d0-b58656ad2a59"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod4d31ee82-2212-4b7f-a4d0-b58656ad2a59] : Timed out while waiting for systemd to remove kubepods-besteffort-pod4d31ee82_2212_4b7f_a4d0_b58656ad2a59.slice" Nov 24 19:33:53 crc kubenswrapper[5035]: I1124 19:33:53.210717 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.118406 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.667348 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-6dq77"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.668694 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.675843 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6dq77"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.831814 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-operator-scripts\") pod \"nova-api-db-create-6dq77\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.831863 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77d9l\" (UniqueName: \"kubernetes.io/projected/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-kube-api-access-77d9l\") pod \"nova-api-db-create-6dq77\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.865630 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-7s5zw"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.866659 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.891746 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6f8f-account-create-57tjg"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.898320 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.902637 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.903786 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7s5zw"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.919416 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6f8f-account-create-57tjg"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.934475 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-operator-scripts\") pod \"nova-api-db-create-6dq77\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.934529 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77d9l\" (UniqueName: \"kubernetes.io/projected/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-kube-api-access-77d9l\") pod \"nova-api-db-create-6dq77\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.935193 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-operator-scripts\") pod \"nova-api-db-create-6dq77\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.970407 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77d9l\" (UniqueName: \"kubernetes.io/projected/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-kube-api-access-77d9l\") pod \"nova-api-db-create-6dq77\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.979773 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-w4w56"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.980959 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.989110 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-w4w56"] Nov 24 19:33:55 crc kubenswrapper[5035]: I1124 19:33:55.997748 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.036213 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec91d4e-a697-46e8-b00e-b7dab491189d-operator-scripts\") pod \"nova-api-6f8f-account-create-57tjg\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.036329 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7m4d\" (UniqueName: \"kubernetes.io/projected/d88bc46f-ae00-4877-b99c-478f8891e14e-kube-api-access-h7m4d\") pod \"nova-cell0-db-create-7s5zw\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.036354 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-247t5\" (UniqueName: \"kubernetes.io/projected/3ec91d4e-a697-46e8-b00e-b7dab491189d-kube-api-access-247t5\") pod \"nova-api-6f8f-account-create-57tjg\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.036396 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88bc46f-ae00-4877-b99c-478f8891e14e-operator-scripts\") pod \"nova-cell0-db-create-7s5zw\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.081614 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-6e5c-account-create-vkrb7"] Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.082844 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.085669 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.097386 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6e5c-account-create-vkrb7"] Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.138113 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88bc46f-ae00-4877-b99c-478f8891e14e-operator-scripts\") pod \"nova-cell0-db-create-7s5zw\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.138612 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgdc2\" (UniqueName: \"kubernetes.io/projected/11977f76-cda4-437d-8ced-e061a8471b5b-kube-api-access-bgdc2\") pod \"nova-cell1-db-create-w4w56\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.138662 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec91d4e-a697-46e8-b00e-b7dab491189d-operator-scripts\") pod \"nova-api-6f8f-account-create-57tjg\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.138737 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11977f76-cda4-437d-8ced-e061a8471b5b-operator-scripts\") pod \"nova-cell1-db-create-w4w56\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.138761 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7m4d\" (UniqueName: \"kubernetes.io/projected/d88bc46f-ae00-4877-b99c-478f8891e14e-kube-api-access-h7m4d\") pod \"nova-cell0-db-create-7s5zw\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.138834 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-247t5\" (UniqueName: \"kubernetes.io/projected/3ec91d4e-a697-46e8-b00e-b7dab491189d-kube-api-access-247t5\") pod \"nova-api-6f8f-account-create-57tjg\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.140019 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88bc46f-ae00-4877-b99c-478f8891e14e-operator-scripts\") pod \"nova-cell0-db-create-7s5zw\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.140603 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec91d4e-a697-46e8-b00e-b7dab491189d-operator-scripts\") pod \"nova-api-6f8f-account-create-57tjg\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.165160 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.165507 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-central-agent" containerID="cri-o://a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76" gracePeriod=30 Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.165613 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="proxy-httpd" containerID="cri-o://62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d" gracePeriod=30 Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.165651 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="sg-core" containerID="cri-o://b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76" gracePeriod=30 Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.165682 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-notification-agent" containerID="cri-o://a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8" gracePeriod=30 Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.168547 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7m4d\" (UniqueName: \"kubernetes.io/projected/d88bc46f-ae00-4877-b99c-478f8891e14e-kube-api-access-h7m4d\") pod \"nova-cell0-db-create-7s5zw\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.183318 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-247t5\" (UniqueName: \"kubernetes.io/projected/3ec91d4e-a697-46e8-b00e-b7dab491189d-kube-api-access-247t5\") pod \"nova-api-6f8f-account-create-57tjg\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.197667 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.215378 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.240558 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l76d7\" (UniqueName: \"kubernetes.io/projected/b75a4d05-f082-474e-b25f-f5889ed418b4-kube-api-access-l76d7\") pod \"nova-cell0-6e5c-account-create-vkrb7\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.240623 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgdc2\" (UniqueName: \"kubernetes.io/projected/11977f76-cda4-437d-8ced-e061a8471b5b-kube-api-access-bgdc2\") pod \"nova-cell1-db-create-w4w56\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.240758 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11977f76-cda4-437d-8ced-e061a8471b5b-operator-scripts\") pod \"nova-cell1-db-create-w4w56\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.240849 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b75a4d05-f082-474e-b25f-f5889ed418b4-operator-scripts\") pod \"nova-cell0-6e5c-account-create-vkrb7\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.242273 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11977f76-cda4-437d-8ced-e061a8471b5b-operator-scripts\") pod \"nova-cell1-db-create-w4w56\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.267532 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgdc2\" (UniqueName: \"kubernetes.io/projected/11977f76-cda4-437d-8ced-e061a8471b5b-kube-api-access-bgdc2\") pod \"nova-cell1-db-create-w4w56\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.300803 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-6317-account-create-j892b"] Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.302268 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.310396 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.314804 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6317-account-create-j892b"] Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.331179 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.344174 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b75a4d05-f082-474e-b25f-f5889ed418b4-operator-scripts\") pod \"nova-cell0-6e5c-account-create-vkrb7\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.344251 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l76d7\" (UniqueName: \"kubernetes.io/projected/b75a4d05-f082-474e-b25f-f5889ed418b4-kube-api-access-l76d7\") pod \"nova-cell0-6e5c-account-create-vkrb7\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.345876 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b75a4d05-f082-474e-b25f-f5889ed418b4-operator-scripts\") pod \"nova-cell0-6e5c-account-create-vkrb7\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.365392 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l76d7\" (UniqueName: \"kubernetes.io/projected/b75a4d05-f082-474e-b25f-f5889ed418b4-kube-api-access-l76d7\") pod \"nova-cell0-6e5c-account-create-vkrb7\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.445835 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0760f5c2-67dd-43ee-918c-75156c940da8-operator-scripts\") pod \"nova-cell1-6317-account-create-j892b\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.445940 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6xdl\" (UniqueName: \"kubernetes.io/projected/0760f5c2-67dd-43ee-918c-75156c940da8-kube-api-access-h6xdl\") pod \"nova-cell1-6317-account-create-j892b\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.547333 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6xdl\" (UniqueName: \"kubernetes.io/projected/0760f5c2-67dd-43ee-918c-75156c940da8-kube-api-access-h6xdl\") pod \"nova-cell1-6317-account-create-j892b\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.547843 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0760f5c2-67dd-43ee-918c-75156c940da8-operator-scripts\") pod \"nova-cell1-6317-account-create-j892b\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.548725 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0760f5c2-67dd-43ee-918c-75156c940da8-operator-scripts\") pod \"nova-cell1-6317-account-create-j892b\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.553917 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.573162 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6xdl\" (UniqueName: \"kubernetes.io/projected/0760f5c2-67dd-43ee-918c-75156c940da8-kube-api-access-h6xdl\") pod \"nova-cell1-6317-account-create-j892b\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.607080 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6dq77"] Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.655625 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.723522 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6ff6c75f76-ttb9x" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.723667 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.898384 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7s5zw"] Nov 24 19:33:56 crc kubenswrapper[5035]: I1124 19:33:56.978464 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6f8f-account-create-57tjg"] Nov 24 19:33:56 crc kubenswrapper[5035]: W1124 19:33:56.987241 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ec91d4e_a697_46e8_b00e_b7dab491189d.slice/crio-a789499d3ee0703d36c59e0819b050ba0ff03eb2d48ad253f616309957e740f9 WatchSource:0}: Error finding container a789499d3ee0703d36c59e0819b050ba0ff03eb2d48ad253f616309957e740f9: Status 404 returned error can't find the container with id a789499d3ee0703d36c59e0819b050ba0ff03eb2d48ad253f616309957e740f9 Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.043713 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6f8f-account-create-57tjg" event={"ID":"3ec91d4e-a697-46e8-b00e-b7dab491189d","Type":"ContainerStarted","Data":"a789499d3ee0703d36c59e0819b050ba0ff03eb2d48ad253f616309957e740f9"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.045678 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dq77" event={"ID":"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa","Type":"ContainerStarted","Data":"aaf5e478c74d73a669b311da730f60aedba30b4a400ba8647986584232a6ee5c"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.045728 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dq77" event={"ID":"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa","Type":"ContainerStarted","Data":"a765e0f438949e5a237af9f4baddba0a75b9297e20e8cda0f9dc3631d2d131af"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.046854 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7s5zw" event={"ID":"d88bc46f-ae00-4877-b99c-478f8891e14e","Type":"ContainerStarted","Data":"7b4ebc8d0477c0d8f6a6ac3e1c100763f732f41f9fb37a132d76a6be0272d119"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.050597 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1128a14c-1553-433f-b64b-676f3cec11f9","Type":"ContainerStarted","Data":"4a476941cc400f3c31fcd7e3390d64ba13a776b83b38798f30acd869893d28d2"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.059229 5035 generic.go:334] "Generic (PLEG): container finished" podID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerID="62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d" exitCode=0 Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.059254 5035 generic.go:334] "Generic (PLEG): container finished" podID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerID="b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76" exitCode=2 Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.059265 5035 generic.go:334] "Generic (PLEG): container finished" podID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerID="a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76" exitCode=0 Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.059299 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerDied","Data":"62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.059340 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerDied","Data":"b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.059371 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerDied","Data":"a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76"} Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.066834 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-6dq77" podStartSLOduration=2.066811472 podStartE2EDuration="2.066811472s" podCreationTimestamp="2025-11-24 19:33:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:33:57.059516444 +0000 UTC m=+1175.582022721" watchObservedRunningTime="2025-11-24 19:33:57.066811472 +0000 UTC m=+1175.589317729" Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.084405 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.490314621 podStartE2EDuration="14.084389299s" podCreationTimestamp="2025-11-24 19:33:43 +0000 UTC" firstStartedPulling="2025-11-24 19:33:44.835150601 +0000 UTC m=+1163.357656858" lastFinishedPulling="2025-11-24 19:33:56.429225279 +0000 UTC m=+1174.951731536" observedRunningTime="2025-11-24 19:33:57.083513275 +0000 UTC m=+1175.606019532" watchObservedRunningTime="2025-11-24 19:33:57.084389299 +0000 UTC m=+1175.606895556" Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.101911 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-w4w56"] Nov 24 19:33:57 crc kubenswrapper[5035]: W1124 19:33:57.129373 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11977f76_cda4_437d_8ced_e061a8471b5b.slice/crio-e816803a500188a87e3afe8d904833a2c378f91a6c17da83ea59c291b23360a6 WatchSource:0}: Error finding container e816803a500188a87e3afe8d904833a2c378f91a6c17da83ea59c291b23360a6: Status 404 returned error can't find the container with id e816803a500188a87e3afe8d904833a2c378f91a6c17da83ea59c291b23360a6 Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.241999 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6317-account-create-j892b"] Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.250333 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6e5c-account-create-vkrb7"] Nov 24 19:33:57 crc kubenswrapper[5035]: W1124 19:33:57.293596 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb75a4d05_f082_474e_b25f_f5889ed418b4.slice/crio-aa51ee5197e94ac6f0d5f55d7b1f7bcb1b7535cb7c4f4c7d472f880a8bd84702 WatchSource:0}: Error finding container aa51ee5197e94ac6f0d5f55d7b1f7bcb1b7535cb7c4f4c7d472f880a8bd84702: Status 404 returned error can't find the container with id aa51ee5197e94ac6f0d5f55d7b1f7bcb1b7535cb7c4f4c7d472f880a8bd84702 Nov 24 19:33:57 crc kubenswrapper[5035]: W1124 19:33:57.294467 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0760f5c2_67dd_43ee_918c_75156c940da8.slice/crio-9d99d55e83571513cdf8a4d90ac4a11757f02eee3475cde40c52ea670f4ec97a WatchSource:0}: Error finding container 9d99d55e83571513cdf8a4d90ac4a11757f02eee3475cde40c52ea670f4ec97a: Status 404 returned error can't find the container with id 9d99d55e83571513cdf8a4d90ac4a11757f02eee3475cde40c52ea670f4ec97a Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.306576 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6b7c585949-9h76s" Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.392646 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6f878cb596-jgfmj"] Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.393331 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6f878cb596-jgfmj" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-httpd" containerID="cri-o://5903b708b5f8b8569216d549b4ed3b3a0261ee7bfab5325abb5b93eedd2751e8" gracePeriod=30 Nov 24 19:33:57 crc kubenswrapper[5035]: I1124 19:33:57.392953 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6f878cb596-jgfmj" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-api" containerID="cri-o://6238cf869c5881652c52e47d931c34742b8e65e50c69f4130ea134e344a5d74a" gracePeriod=30 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.071046 5035 generic.go:334] "Generic (PLEG): container finished" podID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerID="5903b708b5f8b8569216d549b4ed3b3a0261ee7bfab5325abb5b93eedd2751e8" exitCode=0 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.071099 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f878cb596-jgfmj" event={"ID":"4bb5614e-a370-47a0-8cf8-2f4fec8f8537","Type":"ContainerDied","Data":"5903b708b5f8b8569216d549b4ed3b3a0261ee7bfab5325abb5b93eedd2751e8"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.073752 5035 generic.go:334] "Generic (PLEG): container finished" podID="0760f5c2-67dd-43ee-918c-75156c940da8" containerID="0db161e8e9b3e1ee285b374726011b12f15ed9a9b9ead3ef235b3170a01c0354" exitCode=0 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.073815 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6317-account-create-j892b" event={"ID":"0760f5c2-67dd-43ee-918c-75156c940da8","Type":"ContainerDied","Data":"0db161e8e9b3e1ee285b374726011b12f15ed9a9b9ead3ef235b3170a01c0354"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.073861 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6317-account-create-j892b" event={"ID":"0760f5c2-67dd-43ee-918c-75156c940da8","Type":"ContainerStarted","Data":"9d99d55e83571513cdf8a4d90ac4a11757f02eee3475cde40c52ea670f4ec97a"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.075367 5035 generic.go:334] "Generic (PLEG): container finished" podID="11977f76-cda4-437d-8ced-e061a8471b5b" containerID="c7139be1d7916cf66d73e73bde008e314e9f05968ed9a5cdb16dd5962fe4855a" exitCode=0 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.075456 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w4w56" event={"ID":"11977f76-cda4-437d-8ced-e061a8471b5b","Type":"ContainerDied","Data":"c7139be1d7916cf66d73e73bde008e314e9f05968ed9a5cdb16dd5962fe4855a"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.075511 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w4w56" event={"ID":"11977f76-cda4-437d-8ced-e061a8471b5b","Type":"ContainerStarted","Data":"e816803a500188a87e3afe8d904833a2c378f91a6c17da83ea59c291b23360a6"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.078354 5035 generic.go:334] "Generic (PLEG): container finished" podID="b75a4d05-f082-474e-b25f-f5889ed418b4" containerID="8b59b43ccd6101b25ddc3807830175ef03b74064e8932358c6f36b5441c3bde6" exitCode=0 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.078460 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6e5c-account-create-vkrb7" event={"ID":"b75a4d05-f082-474e-b25f-f5889ed418b4","Type":"ContainerDied","Data":"8b59b43ccd6101b25ddc3807830175ef03b74064e8932358c6f36b5441c3bde6"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.078512 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6e5c-account-create-vkrb7" event={"ID":"b75a4d05-f082-474e-b25f-f5889ed418b4","Type":"ContainerStarted","Data":"aa51ee5197e94ac6f0d5f55d7b1f7bcb1b7535cb7c4f4c7d472f880a8bd84702"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.080343 5035 generic.go:334] "Generic (PLEG): container finished" podID="3ec91d4e-a697-46e8-b00e-b7dab491189d" containerID="5c823d874fa8a07d068941d4574987a0b14864ec059594ce6afda6fc58ad9cf2" exitCode=0 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.080420 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6f8f-account-create-57tjg" event={"ID":"3ec91d4e-a697-46e8-b00e-b7dab491189d","Type":"ContainerDied","Data":"5c823d874fa8a07d068941d4574987a0b14864ec059594ce6afda6fc58ad9cf2"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.082521 5035 generic.go:334] "Generic (PLEG): container finished" podID="0e427ed7-2510-4cf8-b9f2-bb30afa79eaa" containerID="aaf5e478c74d73a669b311da730f60aedba30b4a400ba8647986584232a6ee5c" exitCode=0 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.082556 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dq77" event={"ID":"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa","Type":"ContainerDied","Data":"aaf5e478c74d73a669b311da730f60aedba30b4a400ba8647986584232a6ee5c"} Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.084530 5035 generic.go:334] "Generic (PLEG): container finished" podID="d88bc46f-ae00-4877-b99c-478f8891e14e" containerID="f4bc7e2ab83e1721cb4ceba99420481052776e7ed7c3515ce301ef7cf0cc6bb6" exitCode=0 Nov 24 19:33:58 crc kubenswrapper[5035]: I1124 19:33:58.084613 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7s5zw" event={"ID":"d88bc46f-ae00-4877-b99c-478f8891e14e","Type":"ContainerDied","Data":"f4bc7e2ab83e1721cb4ceba99420481052776e7ed7c3515ce301ef7cf0cc6bb6"} Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.550062 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.719721 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0760f5c2-67dd-43ee-918c-75156c940da8-operator-scripts\") pod \"0760f5c2-67dd-43ee-918c-75156c940da8\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.720021 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6xdl\" (UniqueName: \"kubernetes.io/projected/0760f5c2-67dd-43ee-918c-75156c940da8-kube-api-access-h6xdl\") pod \"0760f5c2-67dd-43ee-918c-75156c940da8\" (UID: \"0760f5c2-67dd-43ee-918c-75156c940da8\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.720442 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0760f5c2-67dd-43ee-918c-75156c940da8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0760f5c2-67dd-43ee-918c-75156c940da8" (UID: "0760f5c2-67dd-43ee-918c-75156c940da8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.721516 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0760f5c2-67dd-43ee-918c-75156c940da8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.738582 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0760f5c2-67dd-43ee-918c-75156c940da8-kube-api-access-h6xdl" (OuterVolumeSpecName: "kube-api-access-h6xdl") pod "0760f5c2-67dd-43ee-918c-75156c940da8" (UID: "0760f5c2-67dd-43ee-918c-75156c940da8"). InnerVolumeSpecName "kube-api-access-h6xdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.818467 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dq77" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.822753 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6xdl\" (UniqueName: \"kubernetes.io/projected/0760f5c2-67dd-43ee-918c-75156c940da8-kube-api-access-h6xdl\") on node \"crc\" DevicePath \"\"" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.827578 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.834466 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.845324 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.863772 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.923885 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7m4d\" (UniqueName: \"kubernetes.io/projected/d88bc46f-ae00-4877-b99c-478f8891e14e-kube-api-access-h7m4d\") pod \"d88bc46f-ae00-4877-b99c-478f8891e14e\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.924229 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b75a4d05-f082-474e-b25f-f5889ed418b4-operator-scripts\") pod \"b75a4d05-f082-474e-b25f-f5889ed418b4\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.924337 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgdc2\" (UniqueName: \"kubernetes.io/projected/11977f76-cda4-437d-8ced-e061a8471b5b-kube-api-access-bgdc2\") pod \"11977f76-cda4-437d-8ced-e061a8471b5b\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.924375 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77d9l\" (UniqueName: \"kubernetes.io/projected/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-kube-api-access-77d9l\") pod \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.924422 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l76d7\" (UniqueName: \"kubernetes.io/projected/b75a4d05-f082-474e-b25f-f5889ed418b4-kube-api-access-l76d7\") pod \"b75a4d05-f082-474e-b25f-f5889ed418b4\" (UID: \"b75a4d05-f082-474e-b25f-f5889ed418b4\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.924455 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-operator-scripts\") pod \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\" (UID: \"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.924541 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88bc46f-ae00-4877-b99c-478f8891e14e-operator-scripts\") pod \"d88bc46f-ae00-4877-b99c-478f8891e14e\" (UID: \"d88bc46f-ae00-4877-b99c-478f8891e14e\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.924599 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11977f76-cda4-437d-8ced-e061a8471b5b-operator-scripts\") pod \"11977f76-cda4-437d-8ced-e061a8471b5b\" (UID: \"11977f76-cda4-437d-8ced-e061a8471b5b\") " Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.925417 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11977f76-cda4-437d-8ced-e061a8471b5b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11977f76-cda4-437d-8ced-e061a8471b5b" (UID: "11977f76-cda4-437d-8ced-e061a8471b5b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.925527 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b75a4d05-f082-474e-b25f-f5889ed418b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b75a4d05-f082-474e-b25f-f5889ed418b4" (UID: "b75a4d05-f082-474e-b25f-f5889ed418b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.928793 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d88bc46f-ae00-4877-b99c-478f8891e14e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d88bc46f-ae00-4877-b99c-478f8891e14e" (UID: "d88bc46f-ae00-4877-b99c-478f8891e14e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.929655 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e427ed7-2510-4cf8-b9f2-bb30afa79eaa" (UID: "0e427ed7-2510-4cf8-b9f2-bb30afa79eaa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.929821 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11977f76-cda4-437d-8ced-e061a8471b5b-kube-api-access-bgdc2" (OuterVolumeSpecName: "kube-api-access-bgdc2") pod "11977f76-cda4-437d-8ced-e061a8471b5b" (UID: "11977f76-cda4-437d-8ced-e061a8471b5b"). InnerVolumeSpecName "kube-api-access-bgdc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.929874 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b75a4d05-f082-474e-b25f-f5889ed418b4-kube-api-access-l76d7" (OuterVolumeSpecName: "kube-api-access-l76d7") pod "b75a4d05-f082-474e-b25f-f5889ed418b4" (UID: "b75a4d05-f082-474e-b25f-f5889ed418b4"). InnerVolumeSpecName "kube-api-access-l76d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.930720 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-kube-api-access-77d9l" (OuterVolumeSpecName: "kube-api-access-77d9l") pod "0e427ed7-2510-4cf8-b9f2-bb30afa79eaa" (UID: "0e427ed7-2510-4cf8-b9f2-bb30afa79eaa"). InnerVolumeSpecName "kube-api-access-77d9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:33:59 crc kubenswrapper[5035]: I1124 19:33:59.930839 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d88bc46f-ae00-4877-b99c-478f8891e14e-kube-api-access-h7m4d" (OuterVolumeSpecName: "kube-api-access-h7m4d") pod "d88bc46f-ae00-4877-b99c-478f8891e14e" (UID: "d88bc46f-ae00-4877-b99c-478f8891e14e"). InnerVolumeSpecName "kube-api-access-h7m4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.025966 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-247t5\" (UniqueName: \"kubernetes.io/projected/3ec91d4e-a697-46e8-b00e-b7dab491189d-kube-api-access-247t5\") pod \"3ec91d4e-a697-46e8-b00e-b7dab491189d\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026052 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec91d4e-a697-46e8-b00e-b7dab491189d-operator-scripts\") pod \"3ec91d4e-a697-46e8-b00e-b7dab491189d\" (UID: \"3ec91d4e-a697-46e8-b00e-b7dab491189d\") " Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026363 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11977f76-cda4-437d-8ced-e061a8471b5b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026375 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7m4d\" (UniqueName: \"kubernetes.io/projected/d88bc46f-ae00-4877-b99c-478f8891e14e-kube-api-access-h7m4d\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026385 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b75a4d05-f082-474e-b25f-f5889ed418b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026394 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgdc2\" (UniqueName: \"kubernetes.io/projected/11977f76-cda4-437d-8ced-e061a8471b5b-kube-api-access-bgdc2\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026405 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77d9l\" (UniqueName: \"kubernetes.io/projected/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-kube-api-access-77d9l\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026413 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l76d7\" (UniqueName: \"kubernetes.io/projected/b75a4d05-f082-474e-b25f-f5889ed418b4-kube-api-access-l76d7\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026421 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026429 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88bc46f-ae00-4877-b99c-478f8891e14e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.026625 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ec91d4e-a697-46e8-b00e-b7dab491189d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ec91d4e-a697-46e8-b00e-b7dab491189d" (UID: "3ec91d4e-a697-46e8-b00e-b7dab491189d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.028718 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec91d4e-a697-46e8-b00e-b7dab491189d-kube-api-access-247t5" (OuterVolumeSpecName: "kube-api-access-247t5") pod "3ec91d4e-a697-46e8-b00e-b7dab491189d" (UID: "3ec91d4e-a697-46e8-b00e-b7dab491189d"). InnerVolumeSpecName "kube-api-access-247t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.106980 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6f8f-account-create-57tjg" event={"ID":"3ec91d4e-a697-46e8-b00e-b7dab491189d","Type":"ContainerDied","Data":"a789499d3ee0703d36c59e0819b050ba0ff03eb2d48ad253f616309957e740f9"} Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.107013 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6f8f-account-create-57tjg" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.107026 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a789499d3ee0703d36c59e0819b050ba0ff03eb2d48ad253f616309957e740f9" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.109401 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dq77" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.111621 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dq77" event={"ID":"0e427ed7-2510-4cf8-b9f2-bb30afa79eaa","Type":"ContainerDied","Data":"a765e0f438949e5a237af9f4baddba0a75b9297e20e8cda0f9dc3631d2d131af"} Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.111664 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a765e0f438949e5a237af9f4baddba0a75b9297e20e8cda0f9dc3631d2d131af" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.114260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7s5zw" event={"ID":"d88bc46f-ae00-4877-b99c-478f8891e14e","Type":"ContainerDied","Data":"7b4ebc8d0477c0d8f6a6ac3e1c100763f732f41f9fb37a132d76a6be0272d119"} Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.114286 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b4ebc8d0477c0d8f6a6ac3e1c100763f732f41f9fb37a132d76a6be0272d119" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.114382 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7s5zw" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.117003 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6317-account-create-j892b" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.116979 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6317-account-create-j892b" event={"ID":"0760f5c2-67dd-43ee-918c-75156c940da8","Type":"ContainerDied","Data":"9d99d55e83571513cdf8a4d90ac4a11757f02eee3475cde40c52ea670f4ec97a"} Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.117375 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d99d55e83571513cdf8a4d90ac4a11757f02eee3475cde40c52ea670f4ec97a" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.120473 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4w56" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.120704 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w4w56" event={"ID":"11977f76-cda4-437d-8ced-e061a8471b5b","Type":"ContainerDied","Data":"e816803a500188a87e3afe8d904833a2c378f91a6c17da83ea59c291b23360a6"} Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.120864 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e816803a500188a87e3afe8d904833a2c378f91a6c17da83ea59c291b23360a6" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.124967 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6e5c-account-create-vkrb7" event={"ID":"b75a4d05-f082-474e-b25f-f5889ed418b4","Type":"ContainerDied","Data":"aa51ee5197e94ac6f0d5f55d7b1f7bcb1b7535cb7c4f4c7d472f880a8bd84702"} Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.125007 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa51ee5197e94ac6f0d5f55d7b1f7bcb1b7535cb7c4f4c7d472f880a8bd84702" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.125054 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6e5c-account-create-vkrb7" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.135057 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-247t5\" (UniqueName: \"kubernetes.io/projected/3ec91d4e-a697-46e8-b00e-b7dab491189d-kube-api-access-247t5\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:00 crc kubenswrapper[5035]: I1124 19:34:00.135123 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ec91d4e-a697-46e8-b00e-b7dab491189d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.816434 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969169 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-combined-ca-bundle\") pod \"0b594afb-a6dd-42d4-9430-91e0da323e72\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969246 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-log-httpd\") pod \"0b594afb-a6dd-42d4-9430-91e0da323e72\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969372 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-scripts\") pod \"0b594afb-a6dd-42d4-9430-91e0da323e72\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969408 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-config-data\") pod \"0b594afb-a6dd-42d4-9430-91e0da323e72\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969474 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckzsv\" (UniqueName: \"kubernetes.io/projected/0b594afb-a6dd-42d4-9430-91e0da323e72-kube-api-access-ckzsv\") pod \"0b594afb-a6dd-42d4-9430-91e0da323e72\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969521 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-sg-core-conf-yaml\") pod \"0b594afb-a6dd-42d4-9430-91e0da323e72\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969553 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-run-httpd\") pod \"0b594afb-a6dd-42d4-9430-91e0da323e72\" (UID: \"0b594afb-a6dd-42d4-9430-91e0da323e72\") " Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969820 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0b594afb-a6dd-42d4-9430-91e0da323e72" (UID: "0b594afb-a6dd-42d4-9430-91e0da323e72"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.969883 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0b594afb-a6dd-42d4-9430-91e0da323e72" (UID: "0b594afb-a6dd-42d4-9430-91e0da323e72"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.970227 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.970246 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b594afb-a6dd-42d4-9430-91e0da323e72-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:01 crc kubenswrapper[5035]: I1124 19:34:01.993169 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-scripts" (OuterVolumeSpecName: "scripts") pod "0b594afb-a6dd-42d4-9430-91e0da323e72" (UID: "0b594afb-a6dd-42d4-9430-91e0da323e72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.004103 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b594afb-a6dd-42d4-9430-91e0da323e72-kube-api-access-ckzsv" (OuterVolumeSpecName: "kube-api-access-ckzsv") pod "0b594afb-a6dd-42d4-9430-91e0da323e72" (UID: "0b594afb-a6dd-42d4-9430-91e0da323e72"). InnerVolumeSpecName "kube-api-access-ckzsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.004102 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0b594afb-a6dd-42d4-9430-91e0da323e72" (UID: "0b594afb-a6dd-42d4-9430-91e0da323e72"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.059586 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b594afb-a6dd-42d4-9430-91e0da323e72" (UID: "0b594afb-a6dd-42d4-9430-91e0da323e72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.068272 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-config-data" (OuterVolumeSpecName: "config-data") pod "0b594afb-a6dd-42d4-9430-91e0da323e72" (UID: "0b594afb-a6dd-42d4-9430-91e0da323e72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.071308 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckzsv\" (UniqueName: \"kubernetes.io/projected/0b594afb-a6dd-42d4-9430-91e0da323e72-kube-api-access-ckzsv\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.071339 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.071348 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.071358 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.071369 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b594afb-a6dd-42d4-9430-91e0da323e72-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.143447 5035 generic.go:334] "Generic (PLEG): container finished" podID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerID="a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8" exitCode=0 Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.143492 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerDied","Data":"a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8"} Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.143523 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b594afb-a6dd-42d4-9430-91e0da323e72","Type":"ContainerDied","Data":"7b8921a6b50b4e44f034b26d6791df8cf8f1a30932596ae64f3ccc5bf673024c"} Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.143529 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.143544 5035 scope.go:117] "RemoveContainer" containerID="62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.167655 5035 scope.go:117] "RemoveContainer" containerID="b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.182997 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.192453 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.195864 5035 scope.go:117] "RemoveContainer" containerID="a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.226419 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" path="/var/lib/kubelet/pods/0b594afb-a6dd-42d4-9430-91e0da323e72/volumes" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.230834 5035 scope.go:117] "RemoveContainer" containerID="a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.231650 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.231915 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11977f76-cda4-437d-8ced-e061a8471b5b" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.231931 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="11977f76-cda4-437d-8ced-e061a8471b5b" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.231951 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec91d4e-a697-46e8-b00e-b7dab491189d" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.231957 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec91d4e-a697-46e8-b00e-b7dab491189d" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.231968 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75a4d05-f082-474e-b25f-f5889ed418b4" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.231973 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75a4d05-f082-474e-b25f-f5889ed418b4" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.231986 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-notification-agent" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.231991 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-notification-agent" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.232001 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="proxy-httpd" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232007 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="proxy-httpd" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.232021 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e427ed7-2510-4cf8-b9f2-bb30afa79eaa" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232026 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e427ed7-2510-4cf8-b9f2-bb30afa79eaa" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.232034 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0760f5c2-67dd-43ee-918c-75156c940da8" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232039 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0760f5c2-67dd-43ee-918c-75156c940da8" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.232050 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="sg-core" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232056 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="sg-core" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.232066 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-central-agent" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232073 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-central-agent" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.232084 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88bc46f-ae00-4877-b99c-478f8891e14e" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232089 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88bc46f-ae00-4877-b99c-478f8891e14e" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232234 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="sg-core" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232243 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="proxy-httpd" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232256 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-central-agent" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232369 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec91d4e-a697-46e8-b00e-b7dab491189d" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232383 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="11977f76-cda4-437d-8ced-e061a8471b5b" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232391 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0760f5c2-67dd-43ee-918c-75156c940da8" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232403 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e427ed7-2510-4cf8-b9f2-bb30afa79eaa" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232418 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b594afb-a6dd-42d4-9430-91e0da323e72" containerName="ceilometer-notification-agent" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232426 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b75a4d05-f082-474e-b25f-f5889ed418b4" containerName="mariadb-account-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.232437 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88bc46f-ae00-4877-b99c-478f8891e14e" containerName="mariadb-database-create" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.234970 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.237076 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.237109 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.237361 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.258903 5035 scope.go:117] "RemoveContainer" containerID="62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.259324 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d\": container with ID starting with 62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d not found: ID does not exist" containerID="62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.259358 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d"} err="failed to get container status \"62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d\": rpc error: code = NotFound desc = could not find container \"62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d\": container with ID starting with 62184eb2f640ce759fc5433e076eae1854af9ecd6219653e1487fb1be8ac7c0d not found: ID does not exist" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.259383 5035 scope.go:117] "RemoveContainer" containerID="b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.259868 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76\": container with ID starting with b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76 not found: ID does not exist" containerID="b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.259903 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76"} err="failed to get container status \"b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76\": rpc error: code = NotFound desc = could not find container \"b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76\": container with ID starting with b58418cccfaea57ca03c071a9b84342f35f3d9c6ed676d9160c70f0de4f56b76 not found: ID does not exist" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.259923 5035 scope.go:117] "RemoveContainer" containerID="a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.260433 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8\": container with ID starting with a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8 not found: ID does not exist" containerID="a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.260474 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8"} err="failed to get container status \"a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8\": rpc error: code = NotFound desc = could not find container \"a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8\": container with ID starting with a29f52dee0d9840d355e1909ff07ec35d58b72a53b2570d5e6e4a39cb6167ec8 not found: ID does not exist" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.260487 5035 scope.go:117] "RemoveContainer" containerID="a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76" Nov 24 19:34:02 crc kubenswrapper[5035]: E1124 19:34:02.262515 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76\": container with ID starting with a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76 not found: ID does not exist" containerID="a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.262549 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76"} err="failed to get container status \"a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76\": rpc error: code = NotFound desc = could not find container \"a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76\": container with ID starting with a2142bdae8225cb18ce9bcff1951b5b25968187759bfe71dc6e53a633233cc76 not found: ID does not exist" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.303279 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.303870 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="cc0b947e-0136-4ccf-8203-21604c41ff7d" containerName="kube-state-metrics" containerID="cri-o://f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f" gracePeriod=30 Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.378188 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhwpx\" (UniqueName: \"kubernetes.io/projected/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-kube-api-access-vhwpx\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.378267 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-run-httpd\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.378500 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.378573 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.378659 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-log-httpd\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.378721 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-scripts\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.378787 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-config-data\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480133 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-config-data\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480308 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhwpx\" (UniqueName: \"kubernetes.io/projected/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-kube-api-access-vhwpx\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480355 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-run-httpd\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480383 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480413 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480455 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-log-httpd\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480477 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-scripts\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.480848 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-run-httpd\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.481115 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-log-httpd\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.484523 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.485063 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-scripts\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.485240 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.485962 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-config-data\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.501948 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhwpx\" (UniqueName: \"kubernetes.io/projected/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-kube-api-access-vhwpx\") pod \"ceilometer-0\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.576088 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.703091 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.792404 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg62h\" (UniqueName: \"kubernetes.io/projected/cc0b947e-0136-4ccf-8203-21604c41ff7d-kube-api-access-sg62h\") pod \"cc0b947e-0136-4ccf-8203-21604c41ff7d\" (UID: \"cc0b947e-0136-4ccf-8203-21604c41ff7d\") " Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.814433 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc0b947e-0136-4ccf-8203-21604c41ff7d-kube-api-access-sg62h" (OuterVolumeSpecName: "kube-api-access-sg62h") pod "cc0b947e-0136-4ccf-8203-21604c41ff7d" (UID: "cc0b947e-0136-4ccf-8203-21604c41ff7d"). InnerVolumeSpecName "kube-api-access-sg62h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:02 crc kubenswrapper[5035]: I1124 19:34:02.895441 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg62h\" (UniqueName: \"kubernetes.io/projected/cc0b947e-0136-4ccf-8203-21604c41ff7d-kube-api-access-sg62h\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.089123 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.157818 5035 generic.go:334] "Generic (PLEG): container finished" podID="e67a302e-93be-4341-a7e1-a52183398050" containerID="df984fe11df0c4b1f22de9b25a3845a96fb7f38e7c96e121c2987ad75c805c5c" exitCode=137 Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.157865 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff6c75f76-ttb9x" event={"ID":"e67a302e-93be-4341-a7e1-a52183398050","Type":"ContainerDied","Data":"df984fe11df0c4b1f22de9b25a3845a96fb7f38e7c96e121c2987ad75c805c5c"} Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.158860 5035 generic.go:334] "Generic (PLEG): container finished" podID="cc0b947e-0136-4ccf-8203-21604c41ff7d" containerID="f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f" exitCode=2 Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.158890 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cc0b947e-0136-4ccf-8203-21604c41ff7d","Type":"ContainerDied","Data":"f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f"} Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.158905 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"cc0b947e-0136-4ccf-8203-21604c41ff7d","Type":"ContainerDied","Data":"75477799fff1b06c66d227e174b2ddc0ed65206bd3f1b3ca09e53c6aed6fa773"} Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.158920 5035 scope.go:117] "RemoveContainer" containerID="f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.159015 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.165354 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerStarted","Data":"7db7840b44a08bd449cd4bd185b03503f04f3fb12ab7ca1e2fde14b2b9b73bd3"} Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.166794 5035 generic.go:334] "Generic (PLEG): container finished" podID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerID="6238cf869c5881652c52e47d931c34742b8e65e50c69f4130ea134e344a5d74a" exitCode=0 Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.166817 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f878cb596-jgfmj" event={"ID":"4bb5614e-a370-47a0-8cf8-2f4fec8f8537","Type":"ContainerDied","Data":"6238cf869c5881652c52e47d931c34742b8e65e50c69f4130ea134e344a5d74a"} Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.215943 5035 scope.go:117] "RemoveContainer" containerID="f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.217421 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:34:03 crc kubenswrapper[5035]: E1124 19:34:03.219486 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f\": container with ID starting with f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f not found: ID does not exist" containerID="f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.219520 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f"} err="failed to get container status \"f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f\": rpc error: code = NotFound desc = could not find container \"f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f\": container with ID starting with f1e6493180c5ba2a138fe9660503ad6582fc1e2a3f9b9219903276f76c7df00f not found: ID does not exist" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.226014 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.230460 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.249838 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:34:03 crc kubenswrapper[5035]: E1124 19:34:03.250270 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0b947e-0136-4ccf-8203-21604c41ff7d" containerName="kube-state-metrics" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.250321 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0b947e-0136-4ccf-8203-21604c41ff7d" containerName="kube-state-metrics" Nov 24 19:34:03 crc kubenswrapper[5035]: E1124 19:34:03.250336 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.250342 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" Nov 24 19:34:03 crc kubenswrapper[5035]: E1124 19:34:03.250353 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon-log" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.250360 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon-log" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.250528 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon-log" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.250538 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0b947e-0136-4ccf-8203-21604c41ff7d" containerName="kube-state-metrics" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.250549 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e67a302e-93be-4341-a7e1-a52183398050" containerName="horizon" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.251159 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.256047 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.259230 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.259524 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.303190 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-config-data\") pod \"e67a302e-93be-4341-a7e1-a52183398050\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.303326 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-secret-key\") pod \"e67a302e-93be-4341-a7e1-a52183398050\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.303344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-scripts\") pod \"e67a302e-93be-4341-a7e1-a52183398050\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.303434 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e67a302e-93be-4341-a7e1-a52183398050-logs\") pod \"e67a302e-93be-4341-a7e1-a52183398050\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.303461 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-combined-ca-bundle\") pod \"e67a302e-93be-4341-a7e1-a52183398050\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.303488 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqllc\" (UniqueName: \"kubernetes.io/projected/e67a302e-93be-4341-a7e1-a52183398050-kube-api-access-jqllc\") pod \"e67a302e-93be-4341-a7e1-a52183398050\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.303504 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-tls-certs\") pod \"e67a302e-93be-4341-a7e1-a52183398050\" (UID: \"e67a302e-93be-4341-a7e1-a52183398050\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.304096 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e67a302e-93be-4341-a7e1-a52183398050-logs" (OuterVolumeSpecName: "logs") pod "e67a302e-93be-4341-a7e1-a52183398050" (UID: "e67a302e-93be-4341-a7e1-a52183398050"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.307682 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e67a302e-93be-4341-a7e1-a52183398050" (UID: "e67a302e-93be-4341-a7e1-a52183398050"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.310969 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e67a302e-93be-4341-a7e1-a52183398050-kube-api-access-jqllc" (OuterVolumeSpecName: "kube-api-access-jqllc") pod "e67a302e-93be-4341-a7e1-a52183398050" (UID: "e67a302e-93be-4341-a7e1-a52183398050"). InnerVolumeSpecName "kube-api-access-jqllc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.326328 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-config-data" (OuterVolumeSpecName: "config-data") pod "e67a302e-93be-4341-a7e1-a52183398050" (UID: "e67a302e-93be-4341-a7e1-a52183398050"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.328312 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e67a302e-93be-4341-a7e1-a52183398050" (UID: "e67a302e-93be-4341-a7e1-a52183398050"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.329474 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-scripts" (OuterVolumeSpecName: "scripts") pod "e67a302e-93be-4341-a7e1-a52183398050" (UID: "e67a302e-93be-4341-a7e1-a52183398050"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.355613 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "e67a302e-93be-4341-a7e1-a52183398050" (UID: "e67a302e-93be-4341-a7e1-a52183398050"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405594 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tczb5\" (UniqueName: \"kubernetes.io/projected/bc86b831-9dce-4436-a420-a2f14364bea4-kube-api-access-tczb5\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405644 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405696 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405753 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405823 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405833 5035 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405845 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e67a302e-93be-4341-a7e1-a52183398050-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405852 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e67a302e-93be-4341-a7e1-a52183398050-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405860 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405868 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqllc\" (UniqueName: \"kubernetes.io/projected/e67a302e-93be-4341-a7e1-a52183398050-kube-api-access-jqllc\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.405877 5035 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e67a302e-93be-4341-a7e1-a52183398050-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.407745 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.507200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tczb5\" (UniqueName: \"kubernetes.io/projected/bc86b831-9dce-4436-a420-a2f14364bea4-kube-api-access-tczb5\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.507527 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.507578 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.507608 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.515409 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.515833 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.516414 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc86b831-9dce-4436-a420-a2f14364bea4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.522346 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tczb5\" (UniqueName: \"kubernetes.io/projected/bc86b831-9dce-4436-a420-a2f14364bea4-kube-api-access-tczb5\") pod \"kube-state-metrics-0\" (UID: \"bc86b831-9dce-4436-a420-a2f14364bea4\") " pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.582706 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.727929 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.823839 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-httpd-config\") pod \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.824258 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs\") pod \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.824323 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4p7g\" (UniqueName: \"kubernetes.io/projected/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-kube-api-access-w4p7g\") pod \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.824428 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-config\") pod \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.824448 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-combined-ca-bundle\") pod \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.829413 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4bb5614e-a370-47a0-8cf8-2f4fec8f8537" (UID: "4bb5614e-a370-47a0-8cf8-2f4fec8f8537"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.837480 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-kube-api-access-w4p7g" (OuterVolumeSpecName: "kube-api-access-w4p7g") pod "4bb5614e-a370-47a0-8cf8-2f4fec8f8537" (UID: "4bb5614e-a370-47a0-8cf8-2f4fec8f8537"). InnerVolumeSpecName "kube-api-access-w4p7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.900844 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bb5614e-a370-47a0-8cf8-2f4fec8f8537" (UID: "4bb5614e-a370-47a0-8cf8-2f4fec8f8537"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.904417 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-config" (OuterVolumeSpecName: "config") pod "4bb5614e-a370-47a0-8cf8-2f4fec8f8537" (UID: "4bb5614e-a370-47a0-8cf8-2f4fec8f8537"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.925547 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4bb5614e-a370-47a0-8cf8-2f4fec8f8537" (UID: "4bb5614e-a370-47a0-8cf8-2f4fec8f8537"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.926395 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs\") pod \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\" (UID: \"4bb5614e-a370-47a0-8cf8-2f4fec8f8537\") " Nov 24 19:34:03 crc kubenswrapper[5035]: W1124 19:34:03.926524 5035 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/4bb5614e-a370-47a0-8cf8-2f4fec8f8537/volumes/kubernetes.io~secret/ovndb-tls-certs Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.926542 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4bb5614e-a370-47a0-8cf8-2f4fec8f8537" (UID: "4bb5614e-a370-47a0-8cf8-2f4fec8f8537"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.926997 5035 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.927104 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4p7g\" (UniqueName: \"kubernetes.io/projected/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-kube-api-access-w4p7g\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.927158 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.927216 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:03 crc kubenswrapper[5035]: I1124 19:34:03.927274 5035 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bb5614e-a370-47a0-8cf8-2f4fec8f8537-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.107092 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 19:34:04 crc kubenswrapper[5035]: W1124 19:34:04.110929 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc86b831_9dce_4436_a420_a2f14364bea4.slice/crio-0c7c8d7e49930d02a5cdecd20b4d5680f3019e06e41f2b0889705c808376afb1 WatchSource:0}: Error finding container 0c7c8d7e49930d02a5cdecd20b4d5680f3019e06e41f2b0889705c808376afb1: Status 404 returned error can't find the container with id 0c7c8d7e49930d02a5cdecd20b4d5680f3019e06e41f2b0889705c808376afb1 Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.178387 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f878cb596-jgfmj" event={"ID":"4bb5614e-a370-47a0-8cf8-2f4fec8f8537","Type":"ContainerDied","Data":"799fe7e3633104ea6e51df226bddf64e6732760c5381fcb03a067fec8ebf0a6a"} Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.178640 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f878cb596-jgfmj" Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.178858 5035 scope.go:117] "RemoveContainer" containerID="5903b708b5f8b8569216d549b4ed3b3a0261ee7bfab5325abb5b93eedd2751e8" Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.181756 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff6c75f76-ttb9x" event={"ID":"e67a302e-93be-4341-a7e1-a52183398050","Type":"ContainerDied","Data":"174a679c2a34a3ba9feea167526f1c5a5b875edfe4b94d83ee1f93abd7389b11"} Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.181828 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff6c75f76-ttb9x" Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.183658 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bc86b831-9dce-4436-a420-a2f14364bea4","Type":"ContainerStarted","Data":"0c7c8d7e49930d02a5cdecd20b4d5680f3019e06e41f2b0889705c808376afb1"} Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.213354 5035 scope.go:117] "RemoveContainer" containerID="6238cf869c5881652c52e47d931c34742b8e65e50c69f4130ea134e344a5d74a" Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.215932 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0b947e-0136-4ccf-8203-21604c41ff7d" path="/var/lib/kubelet/pods/cc0b947e-0136-4ccf-8203-21604c41ff7d/volumes" Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.219523 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerStarted","Data":"7ac36c6f1538286b2d7dc71e2f736abb63fdffd5c7c62644769bdd78691eba3a"} Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.229423 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6f878cb596-jgfmj"] Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.240062 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6f878cb596-jgfmj"] Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.248250 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ff6c75f76-ttb9x"] Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.252720 5035 scope.go:117] "RemoveContainer" containerID="b273b803cecfebabd01ad70881af43d61fe52dfee628924d6091e2d362e9ca4c" Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.256720 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6ff6c75f76-ttb9x"] Nov 24 19:34:04 crc kubenswrapper[5035]: I1124 19:34:04.453714 5035 scope.go:117] "RemoveContainer" containerID="df984fe11df0c4b1f22de9b25a3845a96fb7f38e7c96e121c2987ad75c805c5c" Nov 24 19:34:05 crc kubenswrapper[5035]: I1124 19:34:05.220861 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bc86b831-9dce-4436-a420-a2f14364bea4","Type":"ContainerStarted","Data":"3b3d0868f5cb080dc8551755f54b3dff03b4a8257206319eb16bc4f436470b02"} Nov 24 19:34:05 crc kubenswrapper[5035]: I1124 19:34:05.222590 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 19:34:05 crc kubenswrapper[5035]: I1124 19:34:05.252773 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerStarted","Data":"cebadf291014f352333d323c3b56bdd6295489fd1ca8d5343b93f5472502cdaf"} Nov 24 19:34:05 crc kubenswrapper[5035]: I1124 19:34:05.271428 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.839899234 podStartE2EDuration="2.271399351s" podCreationTimestamp="2025-11-24 19:34:03 +0000 UTC" firstStartedPulling="2025-11-24 19:34:04.113100599 +0000 UTC m=+1182.635606856" lastFinishedPulling="2025-11-24 19:34:04.544600716 +0000 UTC m=+1183.067106973" observedRunningTime="2025-11-24 19:34:05.254070291 +0000 UTC m=+1183.776576548" watchObservedRunningTime="2025-11-24 19:34:05.271399351 +0000 UTC m=+1183.793905618" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.146747 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pwhl5"] Nov 24 19:34:06 crc kubenswrapper[5035]: E1124 19:34:06.147313 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-httpd" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.147329 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-httpd" Nov 24 19:34:06 crc kubenswrapper[5035]: E1124 19:34:06.147363 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-api" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.147370 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-api" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.147515 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-httpd" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.147539 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" containerName="neutron-api" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.148105 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.153018 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.156993 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-sfrfx" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.157485 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.169085 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pwhl5"] Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.228246 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb5614e-a370-47a0-8cf8-2f4fec8f8537" path="/var/lib/kubelet/pods/4bb5614e-a370-47a0-8cf8-2f4fec8f8537/volumes" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.228880 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e67a302e-93be-4341-a7e1-a52183398050" path="/var/lib/kubelet/pods/e67a302e-93be-4341-a7e1-a52183398050/volumes" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.265070 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerStarted","Data":"43ca9e70a7d2c1d67aaefddcf7b32a62d48e28675757e2a9ff76dde6718629c8"} Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.273472 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgrrv\" (UniqueName: \"kubernetes.io/projected/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-kube-api-access-tgrrv\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.273588 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.273640 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-config-data\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.273743 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-scripts\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.375597 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-scripts\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.375691 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgrrv\" (UniqueName: \"kubernetes.io/projected/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-kube-api-access-tgrrv\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.376136 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.376174 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-config-data\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.380855 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.381376 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-config-data\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.385655 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-scripts\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.393767 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgrrv\" (UniqueName: \"kubernetes.io/projected/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-kube-api-access-tgrrv\") pod \"nova-cell0-conductor-db-sync-pwhl5\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.467741 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:06 crc kubenswrapper[5035]: I1124 19:34:06.936352 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pwhl5"] Nov 24 19:34:06 crc kubenswrapper[5035]: W1124 19:34:06.937372 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd77dcfaa_154f_4ed4_9fb3_abc2c6af70eb.slice/crio-86ceebd359b19486036c5ac076b2ac789554646b8f0dd6b61b878bd9a1b3338d WatchSource:0}: Error finding container 86ceebd359b19486036c5ac076b2ac789554646b8f0dd6b61b878bd9a1b3338d: Status 404 returned error can't find the container with id 86ceebd359b19486036c5ac076b2ac789554646b8f0dd6b61b878bd9a1b3338d Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.273467 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" event={"ID":"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb","Type":"ContainerStarted","Data":"86ceebd359b19486036c5ac076b2ac789554646b8f0dd6b61b878bd9a1b3338d"} Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.276042 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerStarted","Data":"fa1026d31fbd7774728635c12491104fb28ff27b06cb4b74c3f103c2d64abad1"} Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.276135 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-central-agent" containerID="cri-o://7ac36c6f1538286b2d7dc71e2f736abb63fdffd5c7c62644769bdd78691eba3a" gracePeriod=30 Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.276219 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-notification-agent" containerID="cri-o://cebadf291014f352333d323c3b56bdd6295489fd1ca8d5343b93f5472502cdaf" gracePeriod=30 Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.276222 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="sg-core" containerID="cri-o://43ca9e70a7d2c1d67aaefddcf7b32a62d48e28675757e2a9ff76dde6718629c8" gracePeriod=30 Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.276220 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.276359 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="proxy-httpd" containerID="cri-o://fa1026d31fbd7774728635c12491104fb28ff27b06cb4b74c3f103c2d64abad1" gracePeriod=30 Nov 24 19:34:07 crc kubenswrapper[5035]: I1124 19:34:07.309229 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.866140313 podStartE2EDuration="5.309211597s" podCreationTimestamp="2025-11-24 19:34:02 +0000 UTC" firstStartedPulling="2025-11-24 19:34:03.10335469 +0000 UTC m=+1181.625860947" lastFinishedPulling="2025-11-24 19:34:06.546425974 +0000 UTC m=+1185.068932231" observedRunningTime="2025-11-24 19:34:07.303221234 +0000 UTC m=+1185.825727491" watchObservedRunningTime="2025-11-24 19:34:07.309211597 +0000 UTC m=+1185.831717854" Nov 24 19:34:08 crc kubenswrapper[5035]: I1124 19:34:08.289888 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerID="fa1026d31fbd7774728635c12491104fb28ff27b06cb4b74c3f103c2d64abad1" exitCode=0 Nov 24 19:34:08 crc kubenswrapper[5035]: I1124 19:34:08.291139 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerID="43ca9e70a7d2c1d67aaefddcf7b32a62d48e28675757e2a9ff76dde6718629c8" exitCode=2 Nov 24 19:34:08 crc kubenswrapper[5035]: I1124 19:34:08.291248 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerID="cebadf291014f352333d323c3b56bdd6295489fd1ca8d5343b93f5472502cdaf" exitCode=0 Nov 24 19:34:08 crc kubenswrapper[5035]: I1124 19:34:08.289938 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerDied","Data":"fa1026d31fbd7774728635c12491104fb28ff27b06cb4b74c3f103c2d64abad1"} Nov 24 19:34:08 crc kubenswrapper[5035]: I1124 19:34:08.291506 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerDied","Data":"43ca9e70a7d2c1d67aaefddcf7b32a62d48e28675757e2a9ff76dde6718629c8"} Nov 24 19:34:08 crc kubenswrapper[5035]: I1124 19:34:08.291606 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerDied","Data":"cebadf291014f352333d323c3b56bdd6295489fd1ca8d5343b93f5472502cdaf"} Nov 24 19:34:12 crc kubenswrapper[5035]: I1124 19:34:12.327303 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerID="7ac36c6f1538286b2d7dc71e2f736abb63fdffd5c7c62644769bdd78691eba3a" exitCode=0 Nov 24 19:34:12 crc kubenswrapper[5035]: I1124 19:34:12.327387 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerDied","Data":"7ac36c6f1538286b2d7dc71e2f736abb63fdffd5c7c62644769bdd78691eba3a"} Nov 24 19:34:13 crc kubenswrapper[5035]: I1124 19:34:13.596143 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 19:34:13 crc kubenswrapper[5035]: I1124 19:34:13.886278 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.016846 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhwpx\" (UniqueName: \"kubernetes.io/projected/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-kube-api-access-vhwpx\") pod \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.016920 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-sg-core-conf-yaml\") pod \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.017020 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-run-httpd\") pod \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.017055 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-scripts\") pod \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.017098 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-log-httpd\") pod \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.017234 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-combined-ca-bundle\") pod \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.017329 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-config-data\") pod \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\" (UID: \"7f351c05-cbc1-4fa8-814a-a0273d15b1c3\") " Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.017767 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7f351c05-cbc1-4fa8-814a-a0273d15b1c3" (UID: "7f351c05-cbc1-4fa8-814a-a0273d15b1c3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.017921 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7f351c05-cbc1-4fa8-814a-a0273d15b1c3" (UID: "7f351c05-cbc1-4fa8-814a-a0273d15b1c3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.021711 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-kube-api-access-vhwpx" (OuterVolumeSpecName: "kube-api-access-vhwpx") pod "7f351c05-cbc1-4fa8-814a-a0273d15b1c3" (UID: "7f351c05-cbc1-4fa8-814a-a0273d15b1c3"). InnerVolumeSpecName "kube-api-access-vhwpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.022325 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-scripts" (OuterVolumeSpecName: "scripts") pod "7f351c05-cbc1-4fa8-814a-a0273d15b1c3" (UID: "7f351c05-cbc1-4fa8-814a-a0273d15b1c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.048666 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7f351c05-cbc1-4fa8-814a-a0273d15b1c3" (UID: "7f351c05-cbc1-4fa8-814a-a0273d15b1c3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.116910 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f351c05-cbc1-4fa8-814a-a0273d15b1c3" (UID: "7f351c05-cbc1-4fa8-814a-a0273d15b1c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.118915 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.118946 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.118955 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.118965 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.118976 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhwpx\" (UniqueName: \"kubernetes.io/projected/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-kube-api-access-vhwpx\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.118985 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.154452 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-config-data" (OuterVolumeSpecName: "config-data") pod "7f351c05-cbc1-4fa8-814a-a0273d15b1c3" (UID: "7f351c05-cbc1-4fa8-814a-a0273d15b1c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.220494 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f351c05-cbc1-4fa8-814a-a0273d15b1c3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.350123 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" event={"ID":"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb","Type":"ContainerStarted","Data":"2509916e3fa226dd2f630b71b795266388a2c1a13c67d492fe0fc77b89e04125"} Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.355536 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f351c05-cbc1-4fa8-814a-a0273d15b1c3","Type":"ContainerDied","Data":"7db7840b44a08bd449cd4bd185b03503f04f3fb12ab7ca1e2fde14b2b9b73bd3"} Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.355594 5035 scope.go:117] "RemoveContainer" containerID="fa1026d31fbd7774728635c12491104fb28ff27b06cb4b74c3f103c2d64abad1" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.355764 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.375881 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" podStartSLOduration=1.745102664 podStartE2EDuration="8.375860286s" podCreationTimestamp="2025-11-24 19:34:06 +0000 UTC" firstStartedPulling="2025-11-24 19:34:06.939716854 +0000 UTC m=+1185.462223111" lastFinishedPulling="2025-11-24 19:34:13.570474476 +0000 UTC m=+1192.092980733" observedRunningTime="2025-11-24 19:34:14.37009906 +0000 UTC m=+1192.892605357" watchObservedRunningTime="2025-11-24 19:34:14.375860286 +0000 UTC m=+1192.898366543" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.387693 5035 scope.go:117] "RemoveContainer" containerID="43ca9e70a7d2c1d67aaefddcf7b32a62d48e28675757e2a9ff76dde6718629c8" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.394412 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.404889 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.424661 5035 scope.go:117] "RemoveContainer" containerID="cebadf291014f352333d323c3b56bdd6295489fd1ca8d5343b93f5472502cdaf" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.435892 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:14 crc kubenswrapper[5035]: E1124 19:34:14.436371 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-central-agent" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436392 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-central-agent" Nov 24 19:34:14 crc kubenswrapper[5035]: E1124 19:34:14.436408 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="proxy-httpd" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436415 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="proxy-httpd" Nov 24 19:34:14 crc kubenswrapper[5035]: E1124 19:34:14.436449 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="sg-core" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436457 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="sg-core" Nov 24 19:34:14 crc kubenswrapper[5035]: E1124 19:34:14.436468 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-notification-agent" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436476 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-notification-agent" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436670 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="proxy-httpd" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436691 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="sg-core" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436706 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-notification-agent" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.436718 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" containerName="ceilometer-central-agent" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.438713 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.441697 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.441839 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.442226 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.443448 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.452589 5035 scope.go:117] "RemoveContainer" containerID="7ac36c6f1538286b2d7dc71e2f736abb63fdffd5c7c62644769bdd78691eba3a" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.525926 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.526063 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.526103 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.526143 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.526228 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-scripts\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.526571 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbmfh\" (UniqueName: \"kubernetes.io/projected/be31b03b-6760-4f6a-a8c1-749f13325cc6-kube-api-access-wbmfh\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.526655 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-config-data\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.526751 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628207 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628244 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628273 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628315 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-scripts\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628378 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbmfh\" (UniqueName: \"kubernetes.io/projected/be31b03b-6760-4f6a-a8c1-749f13325cc6-kube-api-access-wbmfh\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628408 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-config-data\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628452 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.628495 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.629075 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.629160 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.633915 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.634019 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.636761 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.637421 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-scripts\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.638922 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-config-data\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.649305 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbmfh\" (UniqueName: \"kubernetes.io/projected/be31b03b-6760-4f6a-a8c1-749f13325cc6-kube-api-access-wbmfh\") pod \"ceilometer-0\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " pod="openstack/ceilometer-0" Nov 24 19:34:14 crc kubenswrapper[5035]: I1124 19:34:14.765248 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:15 crc kubenswrapper[5035]: I1124 19:34:15.201360 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:15 crc kubenswrapper[5035]: W1124 19:34:15.203083 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe31b03b_6760_4f6a_a8c1_749f13325cc6.slice/crio-a9addd38429db2c0dd4b785ff55cbee58275ee94184d306535a4beec6410532b WatchSource:0}: Error finding container a9addd38429db2c0dd4b785ff55cbee58275ee94184d306535a4beec6410532b: Status 404 returned error can't find the container with id a9addd38429db2c0dd4b785ff55cbee58275ee94184d306535a4beec6410532b Nov 24 19:34:15 crc kubenswrapper[5035]: I1124 19:34:15.369664 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerStarted","Data":"a9addd38429db2c0dd4b785ff55cbee58275ee94184d306535a4beec6410532b"} Nov 24 19:34:16 crc kubenswrapper[5035]: I1124 19:34:16.218851 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f351c05-cbc1-4fa8-814a-a0273d15b1c3" path="/var/lib/kubelet/pods/7f351c05-cbc1-4fa8-814a-a0273d15b1c3/volumes" Nov 24 19:34:16 crc kubenswrapper[5035]: I1124 19:34:16.386309 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerStarted","Data":"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7"} Nov 24 19:34:17 crc kubenswrapper[5035]: I1124 19:34:17.399400 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerStarted","Data":"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45"} Nov 24 19:34:18 crc kubenswrapper[5035]: I1124 19:34:18.409904 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerStarted","Data":"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b"} Nov 24 19:34:19 crc kubenswrapper[5035]: I1124 19:34:19.425577 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerStarted","Data":"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0"} Nov 24 19:34:19 crc kubenswrapper[5035]: I1124 19:34:19.426128 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 19:34:19 crc kubenswrapper[5035]: I1124 19:34:19.463649 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.203309248 podStartE2EDuration="5.46362589s" podCreationTimestamp="2025-11-24 19:34:14 +0000 UTC" firstStartedPulling="2025-11-24 19:34:15.205876464 +0000 UTC m=+1193.728382731" lastFinishedPulling="2025-11-24 19:34:18.466193116 +0000 UTC m=+1196.988699373" observedRunningTime="2025-11-24 19:34:19.456817176 +0000 UTC m=+1197.979323493" watchObservedRunningTime="2025-11-24 19:34:19.46362589 +0000 UTC m=+1197.986132147" Nov 24 19:34:21 crc kubenswrapper[5035]: I1124 19:34:21.620457 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:21 crc kubenswrapper[5035]: I1124 19:34:21.620854 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="sg-core" containerID="cri-o://17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" gracePeriod=30 Nov 24 19:34:21 crc kubenswrapper[5035]: I1124 19:34:21.620898 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-notification-agent" containerID="cri-o://6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" gracePeriod=30 Nov 24 19:34:21 crc kubenswrapper[5035]: I1124 19:34:21.620934 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="proxy-httpd" containerID="cri-o://ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" gracePeriod=30 Nov 24 19:34:21 crc kubenswrapper[5035]: I1124 19:34:21.621763 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-central-agent" containerID="cri-o://5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" gracePeriod=30 Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.364127 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457768 5035 generic.go:334] "Generic (PLEG): container finished" podID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerID="ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" exitCode=0 Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457806 5035 generic.go:334] "Generic (PLEG): container finished" podID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerID="17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" exitCode=2 Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457816 5035 generic.go:334] "Generic (PLEG): container finished" podID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerID="6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" exitCode=0 Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457825 5035 generic.go:334] "Generic (PLEG): container finished" podID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerID="5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" exitCode=0 Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457838 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerDied","Data":"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0"} Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457897 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerDied","Data":"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b"} Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457909 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerDied","Data":"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45"} Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457920 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerDied","Data":"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7"} Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457931 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"be31b03b-6760-4f6a-a8c1-749f13325cc6","Type":"ContainerDied","Data":"a9addd38429db2c0dd4b785ff55cbee58275ee94184d306535a4beec6410532b"} Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457861 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.457947 5035 scope.go:117] "RemoveContainer" containerID="ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.486882 5035 scope.go:117] "RemoveContainer" containerID="17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.487946 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbmfh\" (UniqueName: \"kubernetes.io/projected/be31b03b-6760-4f6a-a8c1-749f13325cc6-kube-api-access-wbmfh\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.488071 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-log-httpd\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.488112 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-combined-ca-bundle\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.488136 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-config-data\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.488159 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-scripts\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.488183 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-ceilometer-tls-certs\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.488356 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-run-httpd\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.488679 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-sg-core-conf-yaml\") pod \"be31b03b-6760-4f6a-a8c1-749f13325cc6\" (UID: \"be31b03b-6760-4f6a-a8c1-749f13325cc6\") " Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.491387 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.491579 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.494282 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be31b03b-6760-4f6a-a8c1-749f13325cc6-kube-api-access-wbmfh" (OuterVolumeSpecName: "kube-api-access-wbmfh") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "kube-api-access-wbmfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.495942 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-scripts" (OuterVolumeSpecName: "scripts") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.516776 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.516790 5035 scope.go:117] "RemoveContainer" containerID="6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.541564 5035 scope.go:117] "RemoveContainer" containerID="5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.549757 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.562735 5035 scope.go:117] "RemoveContainer" containerID="ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.563203 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": container with ID starting with ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0 not found: ID does not exist" containerID="ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.563237 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0"} err="failed to get container status \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": rpc error: code = NotFound desc = could not find container \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": container with ID starting with ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.563263 5035 scope.go:117] "RemoveContainer" containerID="17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.563658 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": container with ID starting with 17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b not found: ID does not exist" containerID="17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.563686 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b"} err="failed to get container status \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": rpc error: code = NotFound desc = could not find container \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": container with ID starting with 17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.563702 5035 scope.go:117] "RemoveContainer" containerID="6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.563997 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": container with ID starting with 6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45 not found: ID does not exist" containerID="6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.564028 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45"} err="failed to get container status \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": rpc error: code = NotFound desc = could not find container \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": container with ID starting with 6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.564045 5035 scope.go:117] "RemoveContainer" containerID="5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.564747 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": container with ID starting with 5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7 not found: ID does not exist" containerID="5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.564779 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7"} err="failed to get container status \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": rpc error: code = NotFound desc = could not find container \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": container with ID starting with 5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.564796 5035 scope.go:117] "RemoveContainer" containerID="ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565055 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0"} err="failed to get container status \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": rpc error: code = NotFound desc = could not find container \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": container with ID starting with ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565080 5035 scope.go:117] "RemoveContainer" containerID="17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565322 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b"} err="failed to get container status \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": rpc error: code = NotFound desc = could not find container \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": container with ID starting with 17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565343 5035 scope.go:117] "RemoveContainer" containerID="6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565636 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45"} err="failed to get container status \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": rpc error: code = NotFound desc = could not find container \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": container with ID starting with 6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565660 5035 scope.go:117] "RemoveContainer" containerID="5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565963 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7"} err="failed to get container status \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": rpc error: code = NotFound desc = could not find container \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": container with ID starting with 5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.565978 5035 scope.go:117] "RemoveContainer" containerID="ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.566214 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0"} err="failed to get container status \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": rpc error: code = NotFound desc = could not find container \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": container with ID starting with ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.566271 5035 scope.go:117] "RemoveContainer" containerID="17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.566646 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b"} err="failed to get container status \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": rpc error: code = NotFound desc = could not find container \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": container with ID starting with 17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.566668 5035 scope.go:117] "RemoveContainer" containerID="6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.566991 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45"} err="failed to get container status \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": rpc error: code = NotFound desc = could not find container \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": container with ID starting with 6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.567036 5035 scope.go:117] "RemoveContainer" containerID="5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.567480 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7"} err="failed to get container status \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": rpc error: code = NotFound desc = could not find container \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": container with ID starting with 5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.567510 5035 scope.go:117] "RemoveContainer" containerID="ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.567795 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0"} err="failed to get container status \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": rpc error: code = NotFound desc = could not find container \"ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0\": container with ID starting with ac1e0c3b9b0dee4e6258c4907addc127c5a937ea9fd43e1c4942dd59403737d0 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.567853 5035 scope.go:117] "RemoveContainer" containerID="17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.568218 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b"} err="failed to get container status \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": rpc error: code = NotFound desc = could not find container \"17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b\": container with ID starting with 17e014e3cba111d1be634a5b38feb5b8f8f1193c369b014f52ade7019c3a213b not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.568252 5035 scope.go:117] "RemoveContainer" containerID="6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.568537 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45"} err="failed to get container status \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": rpc error: code = NotFound desc = could not find container \"6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45\": container with ID starting with 6533b3b8d77d2c6b700f773545fa1381cd87c68311bf1025f3ec54856338bc45 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.568575 5035 scope.go:117] "RemoveContainer" containerID="5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.568827 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7"} err="failed to get container status \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": rpc error: code = NotFound desc = could not find container \"5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7\": container with ID starting with 5ff5139cd97845b9caa3176d12af6a2a97ff41589afcf678955c1d759160dcc7 not found: ID does not exist" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.569966 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-config-data" (OuterVolumeSpecName: "config-data") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.570357 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be31b03b-6760-4f6a-a8c1-749f13325cc6" (UID: "be31b03b-6760-4f6a-a8c1-749f13325cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590537 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590559 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590567 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590576 5035 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590584 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590592 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/be31b03b-6760-4f6a-a8c1-749f13325cc6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590600 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbmfh\" (UniqueName: \"kubernetes.io/projected/be31b03b-6760-4f6a-a8c1-749f13325cc6-kube-api-access-wbmfh\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.590609 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/be31b03b-6760-4f6a-a8c1-749f13325cc6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.820498 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.829887 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.840457 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.840873 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-central-agent" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.840894 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-central-agent" Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.840906 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="proxy-httpd" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.840915 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="proxy-httpd" Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.840935 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="sg-core" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.840944 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="sg-core" Nov 24 19:34:22 crc kubenswrapper[5035]: E1124 19:34:22.840970 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-notification-agent" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.840979 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-notification-agent" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.841234 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="sg-core" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.841257 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-notification-agent" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.841274 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="proxy-httpd" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.841326 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" containerName="ceilometer-central-agent" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.843335 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.845153 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.848039 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.848405 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.852533 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.995491 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.995544 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-log-httpd\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.995652 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4crfh\" (UniqueName: \"kubernetes.io/projected/2554287e-1a71-419a-95c0-260b69535239-kube-api-access-4crfh\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.995787 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-config-data\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.995841 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-scripts\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.995890 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.995916 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-run-httpd\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:22 crc kubenswrapper[5035]: I1124 19:34:22.996105 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.101583 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-config-data\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.101746 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-scripts\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.101853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.101926 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-run-httpd\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.102235 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.102357 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.102430 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-log-httpd\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.102537 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4crfh\" (UniqueName: \"kubernetes.io/projected/2554287e-1a71-419a-95c0-260b69535239-kube-api-access-4crfh\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.111444 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-run-httpd\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.111449 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-config-data\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.111998 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-log-httpd\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.117915 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.124974 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-scripts\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.126845 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.128392 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4crfh\" (UniqueName: \"kubernetes.io/projected/2554287e-1a71-419a-95c0-260b69535239-kube-api-access-4crfh\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.128551 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.165957 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:34:23 crc kubenswrapper[5035]: I1124 19:34:23.630473 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:34:23 crc kubenswrapper[5035]: W1124 19:34:23.634351 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2554287e_1a71_419a_95c0_260b69535239.slice/crio-e33563dfd2e612766741e01f83337055e1d71bc85450b8494badd9d4060b7879 WatchSource:0}: Error finding container e33563dfd2e612766741e01f83337055e1d71bc85450b8494badd9d4060b7879: Status 404 returned error can't find the container with id e33563dfd2e612766741e01f83337055e1d71bc85450b8494badd9d4060b7879 Nov 24 19:34:24 crc kubenswrapper[5035]: I1124 19:34:24.217195 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be31b03b-6760-4f6a-a8c1-749f13325cc6" path="/var/lib/kubelet/pods/be31b03b-6760-4f6a-a8c1-749f13325cc6/volumes" Nov 24 19:34:24 crc kubenswrapper[5035]: I1124 19:34:24.480543 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerStarted","Data":"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397"} Nov 24 19:34:24 crc kubenswrapper[5035]: I1124 19:34:24.480597 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerStarted","Data":"e33563dfd2e612766741e01f83337055e1d71bc85450b8494badd9d4060b7879"} Nov 24 19:34:25 crc kubenswrapper[5035]: I1124 19:34:25.494535 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerStarted","Data":"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab"} Nov 24 19:34:26 crc kubenswrapper[5035]: I1124 19:34:26.505328 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerStarted","Data":"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53"} Nov 24 19:34:26 crc kubenswrapper[5035]: I1124 19:34:26.507159 5035 generic.go:334] "Generic (PLEG): container finished" podID="d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" containerID="2509916e3fa226dd2f630b71b795266388a2c1a13c67d492fe0fc77b89e04125" exitCode=0 Nov 24 19:34:26 crc kubenswrapper[5035]: I1124 19:34:26.507187 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" event={"ID":"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb","Type":"ContainerDied","Data":"2509916e3fa226dd2f630b71b795266388a2c1a13c67d492fe0fc77b89e04125"} Nov 24 19:34:27 crc kubenswrapper[5035]: I1124 19:34:27.887921 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:27 crc kubenswrapper[5035]: I1124 19:34:27.992994 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-combined-ca-bundle\") pod \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " Nov 24 19:34:27 crc kubenswrapper[5035]: I1124 19:34:27.994424 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-config-data\") pod \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " Nov 24 19:34:27 crc kubenswrapper[5035]: I1124 19:34:27.994489 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-scripts\") pod \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " Nov 24 19:34:27 crc kubenswrapper[5035]: I1124 19:34:27.994616 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgrrv\" (UniqueName: \"kubernetes.io/projected/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-kube-api-access-tgrrv\") pod \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\" (UID: \"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb\") " Nov 24 19:34:27 crc kubenswrapper[5035]: I1124 19:34:27.998982 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-scripts" (OuterVolumeSpecName: "scripts") pod "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" (UID: "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.000908 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-kube-api-access-tgrrv" (OuterVolumeSpecName: "kube-api-access-tgrrv") pod "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" (UID: "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb"). InnerVolumeSpecName "kube-api-access-tgrrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.025702 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" (UID: "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.028002 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-config-data" (OuterVolumeSpecName: "config-data") pod "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" (UID: "d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.096543 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.096577 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.096586 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgrrv\" (UniqueName: \"kubernetes.io/projected/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-kube-api-access-tgrrv\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.096596 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.534037 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" event={"ID":"d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb","Type":"ContainerDied","Data":"86ceebd359b19486036c5ac076b2ac789554646b8f0dd6b61b878bd9a1b3338d"} Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.534108 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86ceebd359b19486036c5ac076b2ac789554646b8f0dd6b61b878bd9a1b3338d" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.534326 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pwhl5" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.703875 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 19:34:28 crc kubenswrapper[5035]: E1124 19:34:28.706325 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" containerName="nova-cell0-conductor-db-sync" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.706353 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" containerName="nova-cell0-conductor-db-sync" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.706573 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" containerName="nova-cell0-conductor-db-sync" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.707528 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.711221 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-sfrfx" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.711459 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.715449 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.810654 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/817624af-1239-4775-b5d2-87d0c0b7ef95-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.810740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwdrg\" (UniqueName: \"kubernetes.io/projected/817624af-1239-4775-b5d2-87d0c0b7ef95-kube-api-access-rwdrg\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.810812 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/817624af-1239-4775-b5d2-87d0c0b7ef95-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.912665 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwdrg\" (UniqueName: \"kubernetes.io/projected/817624af-1239-4775-b5d2-87d0c0b7ef95-kube-api-access-rwdrg\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.912786 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/817624af-1239-4775-b5d2-87d0c0b7ef95-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.912853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/817624af-1239-4775-b5d2-87d0c0b7ef95-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.917433 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/817624af-1239-4775-b5d2-87d0c0b7ef95-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.929213 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/817624af-1239-4775-b5d2-87d0c0b7ef95-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:28 crc kubenswrapper[5035]: I1124 19:34:28.935969 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwdrg\" (UniqueName: \"kubernetes.io/projected/817624af-1239-4775-b5d2-87d0c0b7ef95-kube-api-access-rwdrg\") pod \"nova-cell0-conductor-0\" (UID: \"817624af-1239-4775-b5d2-87d0c0b7ef95\") " pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:29 crc kubenswrapper[5035]: I1124 19:34:29.044633 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:29 crc kubenswrapper[5035]: I1124 19:34:29.556454 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 19:34:29 crc kubenswrapper[5035]: W1124 19:34:29.581722 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod817624af_1239_4775_b5d2_87d0c0b7ef95.slice/crio-4a3c710214d7c87ca6c835713a2e7f3b854fd97d40c8604d2078ac26e515f9f0 WatchSource:0}: Error finding container 4a3c710214d7c87ca6c835713a2e7f3b854fd97d40c8604d2078ac26e515f9f0: Status 404 returned error can't find the container with id 4a3c710214d7c87ca6c835713a2e7f3b854fd97d40c8604d2078ac26e515f9f0 Nov 24 19:34:30 crc kubenswrapper[5035]: I1124 19:34:30.562732 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerStarted","Data":"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e"} Nov 24 19:34:30 crc kubenswrapper[5035]: I1124 19:34:30.565033 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 19:34:30 crc kubenswrapper[5035]: I1124 19:34:30.566886 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"817624af-1239-4775-b5d2-87d0c0b7ef95","Type":"ContainerStarted","Data":"743ec116602af99e0b2f75332fb1165ccbea903b721778da545ccc9b8263239f"} Nov 24 19:34:30 crc kubenswrapper[5035]: I1124 19:34:30.567063 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"817624af-1239-4775-b5d2-87d0c0b7ef95","Type":"ContainerStarted","Data":"4a3c710214d7c87ca6c835713a2e7f3b854fd97d40c8604d2078ac26e515f9f0"} Nov 24 19:34:30 crc kubenswrapper[5035]: I1124 19:34:30.567257 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:30 crc kubenswrapper[5035]: I1124 19:34:30.605810 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.233233685 podStartE2EDuration="8.605775997s" podCreationTimestamp="2025-11-24 19:34:22 +0000 UTC" firstStartedPulling="2025-11-24 19:34:23.638900496 +0000 UTC m=+1202.161406753" lastFinishedPulling="2025-11-24 19:34:30.011442798 +0000 UTC m=+1208.533949065" observedRunningTime="2025-11-24 19:34:30.593037841 +0000 UTC m=+1209.115544128" watchObservedRunningTime="2025-11-24 19:34:30.605775997 +0000 UTC m=+1209.128282304" Nov 24 19:34:30 crc kubenswrapper[5035]: I1124 19:34:30.627717 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.627690732 podStartE2EDuration="2.627690732s" podCreationTimestamp="2025-11-24 19:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:30.612609792 +0000 UTC m=+1209.135116099" watchObservedRunningTime="2025-11-24 19:34:30.627690732 +0000 UTC m=+1209.150197019" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.089101 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.597131 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xx7p9"] Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.598663 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.607112 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.609522 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.613154 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xx7p9"] Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.731201 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-scripts\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.731255 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4lc8\" (UniqueName: \"kubernetes.io/projected/ebf6870c-e93e-44db-9059-3ca4b23abab3-kube-api-access-q4lc8\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.731377 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-config-data\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.731401 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.791273 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.794918 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.799619 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.837039 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-scripts\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.843789 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4lc8\" (UniqueName: \"kubernetes.io/projected/ebf6870c-e93e-44db-9059-3ca4b23abab3-kube-api-access-q4lc8\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.846280 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-config-data\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.846353 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.837766 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.900444 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-config-data\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.901043 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-scripts\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.901547 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.937932 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4lc8\" (UniqueName: \"kubernetes.io/projected/ebf6870c-e93e-44db-9059-3ca4b23abab3-kube-api-access-q4lc8\") pod \"nova-cell0-cell-mapping-xx7p9\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.943202 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.950268 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-config-data\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.950408 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qlss\" (UniqueName: \"kubernetes.io/projected/ed049f4d-b338-46f4-aeb5-853d0bba9e87-kube-api-access-4qlss\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.950465 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.962356 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.964206 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.980958 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 19:34:34 crc kubenswrapper[5035]: I1124 19:34:34.984759 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.019357 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.020680 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.025260 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.044046 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.053307 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qlss\" (UniqueName: \"kubernetes.io/projected/ed049f4d-b338-46f4-aeb5-853d0bba9e87-kube-api-access-4qlss\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.053382 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.053422 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-config-data\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.053452 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-logs\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.053486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.053540 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-config-data\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.053589 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsz4g\" (UniqueName: \"kubernetes.io/projected/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-kube-api-access-dsz4g\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.061273 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-config-data\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.063114 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.074221 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.076073 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.076994 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qlss\" (UniqueName: \"kubernetes.io/projected/ed049f4d-b338-46f4-aeb5-853d0bba9e87-kube-api-access-4qlss\") pod \"nova-scheduler-0\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.081589 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.099679 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.150408 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f7bbc55bc-tf5sn"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.154603 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.156208 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsz4g\" (UniqueName: \"kubernetes.io/projected/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-kube-api-access-dsz4g\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.156242 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-logs\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.156272 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdcsn\" (UniqueName: \"kubernetes.io/projected/30ab1877-0558-4e64-8ea3-157b2d699faa-kube-api-access-vdcsn\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.156374 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.156397 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4tww\" (UniqueName: \"kubernetes.io/projected/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-kube-api-access-t4tww\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.161843 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.161909 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-config-data\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.161971 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-logs\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.162044 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-config-data\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.162171 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.162200 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.162836 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-logs\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.162834 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f7bbc55bc-tf5sn"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.165597 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.166914 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-config-data\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.169486 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.183936 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsz4g\" (UniqueName: \"kubernetes.io/projected/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-kube-api-access-dsz4g\") pod \"nova-api-0\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.264870 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-config-data\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.264920 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-nb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265005 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265031 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265090 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-logs\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265139 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdcsn\" (UniqueName: \"kubernetes.io/projected/30ab1877-0558-4e64-8ea3-157b2d699faa-kube-api-access-vdcsn\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265198 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-sb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265234 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-config\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265324 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-dns-svc\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4tww\" (UniqueName: \"kubernetes.io/projected/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-kube-api-access-t4tww\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265377 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.265421 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkdbb\" (UniqueName: \"kubernetes.io/projected/301002c3-71f8-41bc-bf96-babd19cc994c-kube-api-access-mkdbb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.266602 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-logs\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.279058 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-config-data\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.279894 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.280588 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.281154 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.282143 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdcsn\" (UniqueName: \"kubernetes.io/projected/30ab1877-0558-4e64-8ea3-157b2d699faa-kube-api-access-vdcsn\") pod \"nova-cell1-novncproxy-0\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.282524 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4tww\" (UniqueName: \"kubernetes.io/projected/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-kube-api-access-t4tww\") pod \"nova-metadata-0\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.366775 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-sb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.366828 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-config\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.366872 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-dns-svc\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.366908 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkdbb\" (UniqueName: \"kubernetes.io/projected/301002c3-71f8-41bc-bf96-babd19cc994c-kube-api-access-mkdbb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.366945 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-nb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.368140 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-config\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.368160 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-sb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.368177 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-nb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.368142 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-dns-svc\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.389320 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkdbb\" (UniqueName: \"kubernetes.io/projected/301002c3-71f8-41bc-bf96-babd19cc994c-kube-api-access-mkdbb\") pod \"dnsmasq-dns-f7bbc55bc-tf5sn\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.470236 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.485760 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.506321 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.520035 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.540159 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xx7p9"] Nov 24 19:34:35 crc kubenswrapper[5035]: W1124 19:34:35.550896 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebf6870c_e93e_44db_9059_3ca4b23abab3.slice/crio-06eb070aefd5ba157806639272f87249d965c928c120895efc45e76fd1866844 WatchSource:0}: Error finding container 06eb070aefd5ba157806639272f87249d965c928c120895efc45e76fd1866844: Status 404 returned error can't find the container with id 06eb070aefd5ba157806639272f87249d965c928c120895efc45e76fd1866844 Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.641356 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lhgxj"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.642703 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.645583 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.649054 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.660746 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lhgxj"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.666812 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xx7p9" event={"ID":"ebf6870c-e93e-44db-9059-3ca4b23abab3","Type":"ContainerStarted","Data":"06eb070aefd5ba157806639272f87249d965c928c120895efc45e76fd1866844"} Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.698462 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.785253 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-config-data\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.785325 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbr77\" (UniqueName: \"kubernetes.io/projected/46c96d7f-2de1-4104-9058-e9fb44e3b677-kube-api-access-qbr77\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.785402 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.785452 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-scripts\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.891859 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.892200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-scripts\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.892333 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-config-data\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.892352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbr77\" (UniqueName: \"kubernetes.io/projected/46c96d7f-2de1-4104-9058-e9fb44e3b677-kube-api-access-qbr77\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.895268 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.897442 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-config-data\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.897884 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-scripts\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.914764 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbr77\" (UniqueName: \"kubernetes.io/projected/46c96d7f-2de1-4104-9058-e9fb44e3b677-kube-api-access-qbr77\") pod \"nova-cell1-conductor-db-sync-lhgxj\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:35 crc kubenswrapper[5035]: I1124 19:34:35.989864 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.121791 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:34:36 crc kubenswrapper[5035]: W1124 19:34:36.125277 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30ab1877_0558_4e64_8ea3_157b2d699faa.slice/crio-fe93ab15c638376edb37ed53b9ceff577ad141b96d503e5481fa5eec35e32346 WatchSource:0}: Error finding container fe93ab15c638376edb37ed53b9ceff577ad141b96d503e5481fa5eec35e32346: Status 404 returned error can't find the container with id fe93ab15c638376edb37ed53b9ceff577ad141b96d503e5481fa5eec35e32346 Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.235910 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.253049 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.362686 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f7bbc55bc-tf5sn"] Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.480389 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lhgxj"] Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.676223 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1","Type":"ContainerStarted","Data":"a6da8996f6b4dec38bd2e3c030b6081d9bd39f0ae9e96bf7a6c4ab8dae5e8e02"} Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.677141 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff","Type":"ContainerStarted","Data":"4c881ffc9e4206be6702b410bd965e3d3d4c1cac4a188a53c998b0bdca6a3c22"} Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.678322 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" event={"ID":"301002c3-71f8-41bc-bf96-babd19cc994c","Type":"ContainerStarted","Data":"25eee20c6e6bd20ed7a0d8b07bb828834923fe78b1e11ca3b82b034396d637f4"} Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.680235 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xx7p9" event={"ID":"ebf6870c-e93e-44db-9059-3ca4b23abab3","Type":"ContainerStarted","Data":"65b044e2ccc2eb9a95bc646b33ae055592bc0466d9ec4a29a3eb320809c90321"} Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.684734 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed049f4d-b338-46f4-aeb5-853d0bba9e87","Type":"ContainerStarted","Data":"1d7aa5c972260162d39f9909981a9a04af1281e10bf58946080ae2c8e1094f09"} Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.687202 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30ab1877-0558-4e64-8ea3-157b2d699faa","Type":"ContainerStarted","Data":"fe93ab15c638376edb37ed53b9ceff577ad141b96d503e5481fa5eec35e32346"} Nov 24 19:34:36 crc kubenswrapper[5035]: I1124 19:34:36.701990 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xx7p9" podStartSLOduration=2.701970314 podStartE2EDuration="2.701970314s" podCreationTimestamp="2025-11-24 19:34:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:36.691944742 +0000 UTC m=+1215.214451019" watchObservedRunningTime="2025-11-24 19:34:36.701970314 +0000 UTC m=+1215.224476571" Nov 24 19:34:36 crc kubenswrapper[5035]: W1124 19:34:36.757324 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46c96d7f_2de1_4104_9058_e9fb44e3b677.slice/crio-38292a577f05e40445d7668e913c5a297e88aa43435ae24da4fada37e55a8187 WatchSource:0}: Error finding container 38292a577f05e40445d7668e913c5a297e88aa43435ae24da4fada37e55a8187: Status 404 returned error can't find the container with id 38292a577f05e40445d7668e913c5a297e88aa43435ae24da4fada37e55a8187 Nov 24 19:34:37 crc kubenswrapper[5035]: I1124 19:34:37.697944 5035 generic.go:334] "Generic (PLEG): container finished" podID="301002c3-71f8-41bc-bf96-babd19cc994c" containerID="fb6fa27cb03fbd7e03a0522a819abe8d9ca1d8d85dba3827283dceeb83a16910" exitCode=0 Nov 24 19:34:37 crc kubenswrapper[5035]: I1124 19:34:37.698006 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" event={"ID":"301002c3-71f8-41bc-bf96-babd19cc994c","Type":"ContainerDied","Data":"fb6fa27cb03fbd7e03a0522a819abe8d9ca1d8d85dba3827283dceeb83a16910"} Nov 24 19:34:37 crc kubenswrapper[5035]: I1124 19:34:37.701685 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" event={"ID":"46c96d7f-2de1-4104-9058-e9fb44e3b677","Type":"ContainerStarted","Data":"dea16e9e471c63ab5f4ee56d465ac26464ba9db13439aed60667c20c0f50a9aa"} Nov 24 19:34:37 crc kubenswrapper[5035]: I1124 19:34:37.701754 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" event={"ID":"46c96d7f-2de1-4104-9058-e9fb44e3b677","Type":"ContainerStarted","Data":"38292a577f05e40445d7668e913c5a297e88aa43435ae24da4fada37e55a8187"} Nov 24 19:34:37 crc kubenswrapper[5035]: I1124 19:34:37.751364 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" podStartSLOduration=2.7513432890000002 podStartE2EDuration="2.751343289s" podCreationTimestamp="2025-11-24 19:34:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:37.742840038 +0000 UTC m=+1216.265346295" watchObservedRunningTime="2025-11-24 19:34:37.751343289 +0000 UTC m=+1216.273849546" Nov 24 19:34:38 crc kubenswrapper[5035]: I1124 19:34:38.603099 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:34:38 crc kubenswrapper[5035]: I1124 19:34:38.611846 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.722280 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" event={"ID":"301002c3-71f8-41bc-bf96-babd19cc994c","Type":"ContainerStarted","Data":"8fbb254c0db5b18f0d54fed5e8b6cf87978057bd5cd5afc94c18688ea509368a"} Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.722630 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.725032 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed049f4d-b338-46f4-aeb5-853d0bba9e87","Type":"ContainerStarted","Data":"ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8"} Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.729942 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30ab1877-0558-4e64-8ea3-157b2d699faa","Type":"ContainerStarted","Data":"d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4"} Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.730071 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="30ab1877-0558-4e64-8ea3-157b2d699faa" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4" gracePeriod=30 Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.742842 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1","Type":"ContainerStarted","Data":"4c0696c2b87e0bd41fbaed214a5bf7b1d0a902951f36ae9b00d9a44f93eb64a9"} Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.742888 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1","Type":"ContainerStarted","Data":"47ba28cd2e6da93f421d7f22340bdb55c53f765266a9e90869a0bf7643816343"} Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.745774 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" podStartSLOduration=4.745761166 podStartE2EDuration="4.745761166s" podCreationTimestamp="2025-11-24 19:34:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:39.743910316 +0000 UTC m=+1218.266416573" watchObservedRunningTime="2025-11-24 19:34:39.745761166 +0000 UTC m=+1218.268267423" Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.766247 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.083672459 podStartE2EDuration="5.766228542s" podCreationTimestamp="2025-11-24 19:34:34 +0000 UTC" firstStartedPulling="2025-11-24 19:34:36.270282402 +0000 UTC m=+1214.792788659" lastFinishedPulling="2025-11-24 19:34:38.952838485 +0000 UTC m=+1217.475344742" observedRunningTime="2025-11-24 19:34:39.763744504 +0000 UTC m=+1218.286250761" watchObservedRunningTime="2025-11-24 19:34:39.766228542 +0000 UTC m=+1218.288734799" Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.767482 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff","Type":"ContainerStarted","Data":"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028"} Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.767525 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff","Type":"ContainerStarted","Data":"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5"} Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.767638 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-log" containerID="cri-o://e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5" gracePeriod=30 Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.767760 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-metadata" containerID="cri-o://18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028" gracePeriod=30 Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.788709 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.585182724 podStartE2EDuration="5.788690212s" podCreationTimestamp="2025-11-24 19:34:34 +0000 UTC" firstStartedPulling="2025-11-24 19:34:35.725612032 +0000 UTC m=+1214.248118289" lastFinishedPulling="2025-11-24 19:34:38.92911952 +0000 UTC m=+1217.451625777" observedRunningTime="2025-11-24 19:34:39.780644463 +0000 UTC m=+1218.303150720" watchObservedRunningTime="2025-11-24 19:34:39.788690212 +0000 UTC m=+1218.311196469" Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.799911 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.972936112 podStartE2EDuration="5.799884976s" podCreationTimestamp="2025-11-24 19:34:34 +0000 UTC" firstStartedPulling="2025-11-24 19:34:36.127745551 +0000 UTC m=+1214.650251808" lastFinishedPulling="2025-11-24 19:34:38.954694415 +0000 UTC m=+1217.477200672" observedRunningTime="2025-11-24 19:34:39.794680874 +0000 UTC m=+1218.317187131" watchObservedRunningTime="2025-11-24 19:34:39.799884976 +0000 UTC m=+1218.322391233" Nov 24 19:34:39 crc kubenswrapper[5035]: I1124 19:34:39.827215 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.144339456 podStartE2EDuration="5.827197537s" podCreationTimestamp="2025-11-24 19:34:34 +0000 UTC" firstStartedPulling="2025-11-24 19:34:36.276898302 +0000 UTC m=+1214.799404559" lastFinishedPulling="2025-11-24 19:34:38.959756373 +0000 UTC m=+1217.482262640" observedRunningTime="2025-11-24 19:34:39.819883878 +0000 UTC m=+1218.342390135" watchObservedRunningTime="2025-11-24 19:34:39.827197537 +0000 UTC m=+1218.349703794" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.169412 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.485987 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.506993 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.507059 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.651254 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.777371 5035 generic.go:334] "Generic (PLEG): container finished" podID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerID="18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028" exitCode=0 Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.777400 5035 generic.go:334] "Generic (PLEG): container finished" podID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerID="e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5" exitCode=143 Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.778214 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.778506 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff","Type":"ContainerDied","Data":"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028"} Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.778591 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff","Type":"ContainerDied","Data":"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5"} Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.778611 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff","Type":"ContainerDied","Data":"4c881ffc9e4206be6702b410bd965e3d3d4c1cac4a188a53c998b0bdca6a3c22"} Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.778644 5035 scope.go:117] "RemoveContainer" containerID="18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.801038 5035 scope.go:117] "RemoveContainer" containerID="e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.803183 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-combined-ca-bundle\") pod \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.803239 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-config-data\") pod \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.803332 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-logs\") pod \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.803412 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4tww\" (UniqueName: \"kubernetes.io/projected/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-kube-api-access-t4tww\") pod \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\" (UID: \"fe0474c1-8430-4a19-b2f4-cd5f5c9650ff\") " Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.804018 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-logs" (OuterVolumeSpecName: "logs") pod "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" (UID: "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.804257 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.825716 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-kube-api-access-t4tww" (OuterVolumeSpecName: "kube-api-access-t4tww") pod "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" (UID: "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff"). InnerVolumeSpecName "kube-api-access-t4tww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.829535 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" (UID: "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.842580 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-config-data" (OuterVolumeSpecName: "config-data") pod "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" (UID: "fe0474c1-8430-4a19-b2f4-cd5f5c9650ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.884708 5035 scope.go:117] "RemoveContainer" containerID="18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028" Nov 24 19:34:40 crc kubenswrapper[5035]: E1124 19:34:40.885305 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028\": container with ID starting with 18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028 not found: ID does not exist" containerID="18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.885351 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028"} err="failed to get container status \"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028\": rpc error: code = NotFound desc = could not find container \"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028\": container with ID starting with 18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028 not found: ID does not exist" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.885382 5035 scope.go:117] "RemoveContainer" containerID="e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5" Nov 24 19:34:40 crc kubenswrapper[5035]: E1124 19:34:40.885645 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5\": container with ID starting with e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5 not found: ID does not exist" containerID="e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.885664 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5"} err="failed to get container status \"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5\": rpc error: code = NotFound desc = could not find container \"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5\": container with ID starting with e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5 not found: ID does not exist" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.885676 5035 scope.go:117] "RemoveContainer" containerID="18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.885850 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028"} err="failed to get container status \"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028\": rpc error: code = NotFound desc = could not find container \"18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028\": container with ID starting with 18f2761ebbb5d0fcde894b6b0094de1607e130cbc4a8ae18356b7414214e1028 not found: ID does not exist" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.885863 5035 scope.go:117] "RemoveContainer" containerID="e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.886003 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5"} err="failed to get container status \"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5\": rpc error: code = NotFound desc = could not find container \"e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5\": container with ID starting with e70d8384a5091200864a6ea02a0eb1e302611f64e4a9613c84439ae11a9d11b5 not found: ID does not exist" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.905917 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4tww\" (UniqueName: \"kubernetes.io/projected/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-kube-api-access-t4tww\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.905953 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:40 crc kubenswrapper[5035]: I1124 19:34:40.905964 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.115703 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.124872 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.137585 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:41 crc kubenswrapper[5035]: E1124 19:34:41.138099 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-log" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.138117 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-log" Nov 24 19:34:41 crc kubenswrapper[5035]: E1124 19:34:41.138173 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-metadata" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.138194 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-metadata" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.138521 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-log" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.138540 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" containerName="nova-metadata-metadata" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.140431 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.142962 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.143253 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.151837 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.211521 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4vh\" (UniqueName: \"kubernetes.io/projected/9f00d60b-a125-4768-b123-90859a45671d-kube-api-access-km4vh\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.211626 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.211661 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-config-data\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.211722 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f00d60b-a125-4768-b123-90859a45671d-logs\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.211781 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.313590 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4vh\" (UniqueName: \"kubernetes.io/projected/9f00d60b-a125-4768-b123-90859a45671d-kube-api-access-km4vh\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.313715 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.313760 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-config-data\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.313789 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f00d60b-a125-4768-b123-90859a45671d-logs\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.313851 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.314828 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f00d60b-a125-4768-b123-90859a45671d-logs\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.328691 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.328839 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-config-data\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.329170 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.340741 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4vh\" (UniqueName: \"kubernetes.io/projected/9f00d60b-a125-4768-b123-90859a45671d-kube-api-access-km4vh\") pod \"nova-metadata-0\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.479875 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:41 crc kubenswrapper[5035]: I1124 19:34:41.953049 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:41 crc kubenswrapper[5035]: W1124 19:34:41.953440 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f00d60b_a125_4768_b123_90859a45671d.slice/crio-aea3b5fc7aabde99a901a348567f0d0ff6741b20a5271b9453a3cd356bab4c72 WatchSource:0}: Error finding container aea3b5fc7aabde99a901a348567f0d0ff6741b20a5271b9453a3cd356bab4c72: Status 404 returned error can't find the container with id aea3b5fc7aabde99a901a348567f0d0ff6741b20a5271b9453a3cd356bab4c72 Nov 24 19:34:42 crc kubenswrapper[5035]: I1124 19:34:42.222912 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe0474c1-8430-4a19-b2f4-cd5f5c9650ff" path="/var/lib/kubelet/pods/fe0474c1-8430-4a19-b2f4-cd5f5c9650ff/volumes" Nov 24 19:34:42 crc kubenswrapper[5035]: I1124 19:34:42.800139 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f00d60b-a125-4768-b123-90859a45671d","Type":"ContainerStarted","Data":"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05"} Nov 24 19:34:42 crc kubenswrapper[5035]: I1124 19:34:42.800228 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f00d60b-a125-4768-b123-90859a45671d","Type":"ContainerStarted","Data":"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8"} Nov 24 19:34:42 crc kubenswrapper[5035]: I1124 19:34:42.800260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f00d60b-a125-4768-b123-90859a45671d","Type":"ContainerStarted","Data":"aea3b5fc7aabde99a901a348567f0d0ff6741b20a5271b9453a3cd356bab4c72"} Nov 24 19:34:42 crc kubenswrapper[5035]: I1124 19:34:42.834192 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.834170879 podStartE2EDuration="1.834170879s" podCreationTimestamp="2025-11-24 19:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:42.81873774 +0000 UTC m=+1221.341244027" watchObservedRunningTime="2025-11-24 19:34:42.834170879 +0000 UTC m=+1221.356677146" Nov 24 19:34:43 crc kubenswrapper[5035]: I1124 19:34:43.814802 5035 generic.go:334] "Generic (PLEG): container finished" podID="ebf6870c-e93e-44db-9059-3ca4b23abab3" containerID="65b044e2ccc2eb9a95bc646b33ae055592bc0466d9ec4a29a3eb320809c90321" exitCode=0 Nov 24 19:34:43 crc kubenswrapper[5035]: I1124 19:34:43.814890 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xx7p9" event={"ID":"ebf6870c-e93e-44db-9059-3ca4b23abab3","Type":"ContainerDied","Data":"65b044e2ccc2eb9a95bc646b33ae055592bc0466d9ec4a29a3eb320809c90321"} Nov 24 19:34:44 crc kubenswrapper[5035]: I1124 19:34:44.831170 5035 generic.go:334] "Generic (PLEG): container finished" podID="46c96d7f-2de1-4104-9058-e9fb44e3b677" containerID="dea16e9e471c63ab5f4ee56d465ac26464ba9db13439aed60667c20c0f50a9aa" exitCode=0 Nov 24 19:34:44 crc kubenswrapper[5035]: I1124 19:34:44.831331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" event={"ID":"46c96d7f-2de1-4104-9058-e9fb44e3b677","Type":"ContainerDied","Data":"dea16e9e471c63ab5f4ee56d465ac26464ba9db13439aed60667c20c0f50a9aa"} Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.166791 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.212868 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.272180 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.386937 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4lc8\" (UniqueName: \"kubernetes.io/projected/ebf6870c-e93e-44db-9059-3ca4b23abab3-kube-api-access-q4lc8\") pod \"ebf6870c-e93e-44db-9059-3ca4b23abab3\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.387030 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-combined-ca-bundle\") pod \"ebf6870c-e93e-44db-9059-3ca4b23abab3\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.387383 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-scripts\") pod \"ebf6870c-e93e-44db-9059-3ca4b23abab3\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.387414 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-config-data\") pod \"ebf6870c-e93e-44db-9059-3ca4b23abab3\" (UID: \"ebf6870c-e93e-44db-9059-3ca4b23abab3\") " Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.392721 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf6870c-e93e-44db-9059-3ca4b23abab3-kube-api-access-q4lc8" (OuterVolumeSpecName: "kube-api-access-q4lc8") pod "ebf6870c-e93e-44db-9059-3ca4b23abab3" (UID: "ebf6870c-e93e-44db-9059-3ca4b23abab3"). InnerVolumeSpecName "kube-api-access-q4lc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.393045 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-scripts" (OuterVolumeSpecName: "scripts") pod "ebf6870c-e93e-44db-9059-3ca4b23abab3" (UID: "ebf6870c-e93e-44db-9059-3ca4b23abab3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.414814 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-config-data" (OuterVolumeSpecName: "config-data") pod "ebf6870c-e93e-44db-9059-3ca4b23abab3" (UID: "ebf6870c-e93e-44db-9059-3ca4b23abab3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.429199 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebf6870c-e93e-44db-9059-3ca4b23abab3" (UID: "ebf6870c-e93e-44db-9059-3ca4b23abab3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.471352 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.471419 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.489455 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.489495 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.489507 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4lc8\" (UniqueName: \"kubernetes.io/projected/ebf6870c-e93e-44db-9059-3ca4b23abab3-kube-api-access-q4lc8\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.489517 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebf6870c-e93e-44db-9059-3ca4b23abab3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.523385 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.587151 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bc89f58d7-sc6h8"] Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.587687 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" podUID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerName="dnsmasq-dns" containerID="cri-o://53794eb118a5ab08840029397a1db927b614e05877a303e0e0a52c339c579b93" gracePeriod=10 Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.881378 5035 generic.go:334] "Generic (PLEG): container finished" podID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerID="53794eb118a5ab08840029397a1db927b614e05877a303e0e0a52c339c579b93" exitCode=0 Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.881592 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" event={"ID":"c6cf7f09-7707-4a27-a61e-1d49897c2e63","Type":"ContainerDied","Data":"53794eb118a5ab08840029397a1db927b614e05877a303e0e0a52c339c579b93"} Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.884026 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xx7p9" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.884795 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xx7p9" event={"ID":"ebf6870c-e93e-44db-9059-3ca4b23abab3","Type":"ContainerDied","Data":"06eb070aefd5ba157806639272f87249d965c928c120895efc45e76fd1866844"} Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.884837 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06eb070aefd5ba157806639272f87249d965c928c120895efc45e76fd1866844" Nov 24 19:34:45 crc kubenswrapper[5035]: I1124 19:34:45.942310 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.121938 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.126183 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.126394 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-log" containerID="cri-o://47ba28cd2e6da93f421d7f22340bdb55c53f765266a9e90869a0bf7643816343" gracePeriod=30 Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.126603 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-api" containerID="cri-o://4c0696c2b87e0bd41fbaed214a5bf7b1d0a902951f36ae9b00d9a44f93eb64a9" gracePeriod=30 Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.136778 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.176:8774/\": EOF" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.136791 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.176:8774/\": EOF" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.198504 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.199102 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-log" containerID="cri-o://8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8" gracePeriod=30 Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.199507 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-metadata" containerID="cri-o://babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05" gracePeriod=30 Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.208209 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-dns-svc\") pod \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.208314 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-nb\") pod \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.208444 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-config\") pod \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.208515 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-sb\") pod \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.208606 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxh4m\" (UniqueName: \"kubernetes.io/projected/c6cf7f09-7707-4a27-a61e-1d49897c2e63-kube-api-access-vxh4m\") pod \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\" (UID: \"c6cf7f09-7707-4a27-a61e-1d49897c2e63\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.219517 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6cf7f09-7707-4a27-a61e-1d49897c2e63-kube-api-access-vxh4m" (OuterVolumeSpecName: "kube-api-access-vxh4m") pod "c6cf7f09-7707-4a27-a61e-1d49897c2e63" (UID: "c6cf7f09-7707-4a27-a61e-1d49897c2e63"). InnerVolumeSpecName "kube-api-access-vxh4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.285211 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c6cf7f09-7707-4a27-a61e-1d49897c2e63" (UID: "c6cf7f09-7707-4a27-a61e-1d49897c2e63"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.285400 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c6cf7f09-7707-4a27-a61e-1d49897c2e63" (UID: "c6cf7f09-7707-4a27-a61e-1d49897c2e63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.290790 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c6cf7f09-7707-4a27-a61e-1d49897c2e63" (UID: "c6cf7f09-7707-4a27-a61e-1d49897c2e63"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.291364 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-config" (OuterVolumeSpecName: "config") pod "c6cf7f09-7707-4a27-a61e-1d49897c2e63" (UID: "c6cf7f09-7707-4a27-a61e-1d49897c2e63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.310691 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxh4m\" (UniqueName: \"kubernetes.io/projected/c6cf7f09-7707-4a27-a61e-1d49897c2e63-kube-api-access-vxh4m\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.310722 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.310733 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.310742 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.310750 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c6cf7f09-7707-4a27-a61e-1d49897c2e63-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.457805 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.457939 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.480250 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.480309 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.615997 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-scripts\") pod \"46c96d7f-2de1-4104-9058-e9fb44e3b677\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.616031 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-config-data\") pod \"46c96d7f-2de1-4104-9058-e9fb44e3b677\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.616098 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-combined-ca-bundle\") pod \"46c96d7f-2de1-4104-9058-e9fb44e3b677\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.616225 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbr77\" (UniqueName: \"kubernetes.io/projected/46c96d7f-2de1-4104-9058-e9fb44e3b677-kube-api-access-qbr77\") pod \"46c96d7f-2de1-4104-9058-e9fb44e3b677\" (UID: \"46c96d7f-2de1-4104-9058-e9fb44e3b677\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.626542 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-scripts" (OuterVolumeSpecName: "scripts") pod "46c96d7f-2de1-4104-9058-e9fb44e3b677" (UID: "46c96d7f-2de1-4104-9058-e9fb44e3b677"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.629438 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46c96d7f-2de1-4104-9058-e9fb44e3b677-kube-api-access-qbr77" (OuterVolumeSpecName: "kube-api-access-qbr77") pod "46c96d7f-2de1-4104-9058-e9fb44e3b677" (UID: "46c96d7f-2de1-4104-9058-e9fb44e3b677"). InnerVolumeSpecName "kube-api-access-qbr77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.649603 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46c96d7f-2de1-4104-9058-e9fb44e3b677" (UID: "46c96d7f-2de1-4104-9058-e9fb44e3b677"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.656784 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-config-data" (OuterVolumeSpecName: "config-data") pod "46c96d7f-2de1-4104-9058-e9fb44e3b677" (UID: "46c96d7f-2de1-4104-9058-e9fb44e3b677"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.718237 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.718280 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.718404 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c96d7f-2de1-4104-9058-e9fb44e3b677-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.718422 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbr77\" (UniqueName: \"kubernetes.io/projected/46c96d7f-2de1-4104-9058-e9fb44e3b677-kube-api-access-qbr77\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.726632 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.819830 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-combined-ca-bundle\") pod \"9f00d60b-a125-4768-b123-90859a45671d\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.819895 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-nova-metadata-tls-certs\") pod \"9f00d60b-a125-4768-b123-90859a45671d\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.819990 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-config-data\") pod \"9f00d60b-a125-4768-b123-90859a45671d\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.820012 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km4vh\" (UniqueName: \"kubernetes.io/projected/9f00d60b-a125-4768-b123-90859a45671d-kube-api-access-km4vh\") pod \"9f00d60b-a125-4768-b123-90859a45671d\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.820157 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f00d60b-a125-4768-b123-90859a45671d-logs\") pod \"9f00d60b-a125-4768-b123-90859a45671d\" (UID: \"9f00d60b-a125-4768-b123-90859a45671d\") " Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.820766 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f00d60b-a125-4768-b123-90859a45671d-logs" (OuterVolumeSpecName: "logs") pod "9f00d60b-a125-4768-b123-90859a45671d" (UID: "9f00d60b-a125-4768-b123-90859a45671d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.836841 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f00d60b-a125-4768-b123-90859a45671d-kube-api-access-km4vh" (OuterVolumeSpecName: "kube-api-access-km4vh") pod "9f00d60b-a125-4768-b123-90859a45671d" (UID: "9f00d60b-a125-4768-b123-90859a45671d"). InnerVolumeSpecName "kube-api-access-km4vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.852458 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f00d60b-a125-4768-b123-90859a45671d" (UID: "9f00d60b-a125-4768-b123-90859a45671d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.863162 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-config-data" (OuterVolumeSpecName: "config-data") pod "9f00d60b-a125-4768-b123-90859a45671d" (UID: "9f00d60b-a125-4768-b123-90859a45671d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.895162 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" event={"ID":"46c96d7f-2de1-4104-9058-e9fb44e3b677","Type":"ContainerDied","Data":"38292a577f05e40445d7668e913c5a297e88aa43435ae24da4fada37e55a8187"} Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.896146 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38292a577f05e40445d7668e913c5a297e88aa43435ae24da4fada37e55a8187" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.895447 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lhgxj" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.896986 5035 generic.go:334] "Generic (PLEG): container finished" podID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerID="47ba28cd2e6da93f421d7f22340bdb55c53f765266a9e90869a0bf7643816343" exitCode=143 Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.897078 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1","Type":"ContainerDied","Data":"47ba28cd2e6da93f421d7f22340bdb55c53f765266a9e90869a0bf7643816343"} Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.900099 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.900107 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89f58d7-sc6h8" event={"ID":"c6cf7f09-7707-4a27-a61e-1d49897c2e63","Type":"ContainerDied","Data":"e98a6aef6f8ead4b2f901f27bcaa6a22ae99eb7219ce29ab781ffdc569acfe2a"} Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.900177 5035 scope.go:117] "RemoveContainer" containerID="53794eb118a5ab08840029397a1db927b614e05877a303e0e0a52c339c579b93" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.904047 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9f00d60b-a125-4768-b123-90859a45671d" (UID: "9f00d60b-a125-4768-b123-90859a45671d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.904467 5035 generic.go:334] "Generic (PLEG): container finished" podID="9f00d60b-a125-4768-b123-90859a45671d" containerID="babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05" exitCode=0 Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.904487 5035 generic.go:334] "Generic (PLEG): container finished" podID="9f00d60b-a125-4768-b123-90859a45671d" containerID="8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8" exitCode=143 Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.905460 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.906276 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f00d60b-a125-4768-b123-90859a45671d","Type":"ContainerDied","Data":"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05"} Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.906320 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f00d60b-a125-4768-b123-90859a45671d","Type":"ContainerDied","Data":"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8"} Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.906331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f00d60b-a125-4768-b123-90859a45671d","Type":"ContainerDied","Data":"aea3b5fc7aabde99a901a348567f0d0ff6741b20a5271b9453a3cd356bab4c72"} Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.922433 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f00d60b-a125-4768-b123-90859a45671d-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.922461 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.922472 5035 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.922483 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f00d60b-a125-4768-b123-90859a45671d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.922492 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km4vh\" (UniqueName: \"kubernetes.io/projected/9f00d60b-a125-4768-b123-90859a45671d-kube-api-access-km4vh\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.925044 5035 scope.go:117] "RemoveContainer" containerID="67fe3d6d4e1fa113da4420884d9c72448baa4c96f66f8d353ad7d8790bce4c3a" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.962491 5035 scope.go:117] "RemoveContainer" containerID="babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.967656 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bc89f58d7-sc6h8"] Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.983955 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bc89f58d7-sc6h8"] Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.986507 5035 scope.go:117] "RemoveContainer" containerID="8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.993888 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 19:34:46 crc kubenswrapper[5035]: E1124 19:34:46.994229 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerName="init" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.994245 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerName="init" Nov 24 19:34:46 crc kubenswrapper[5035]: E1124 19:34:46.994264 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerName="dnsmasq-dns" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.994274 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerName="dnsmasq-dns" Nov 24 19:34:46 crc kubenswrapper[5035]: E1124 19:34:46.994311 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf6870c-e93e-44db-9059-3ca4b23abab3" containerName="nova-manage" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.994317 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf6870c-e93e-44db-9059-3ca4b23abab3" containerName="nova-manage" Nov 24 19:34:46 crc kubenswrapper[5035]: E1124 19:34:46.995509 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c96d7f-2de1-4104-9058-e9fb44e3b677" containerName="nova-cell1-conductor-db-sync" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.995537 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c96d7f-2de1-4104-9058-e9fb44e3b677" containerName="nova-cell1-conductor-db-sync" Nov 24 19:34:46 crc kubenswrapper[5035]: E1124 19:34:46.995552 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-metadata" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.995561 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-metadata" Nov 24 19:34:46 crc kubenswrapper[5035]: E1124 19:34:46.995578 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-log" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.995587 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-log" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.995976 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" containerName="dnsmasq-dns" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.996041 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf6870c-e93e-44db-9059-3ca4b23abab3" containerName="nova-manage" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.996088 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-metadata" Nov 24 19:34:46 crc kubenswrapper[5035]: I1124 19:34:46.996104 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f00d60b-a125-4768-b123-90859a45671d" containerName="nova-metadata-log" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.002392 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="46c96d7f-2de1-4104-9058-e9fb44e3b677" containerName="nova-cell1-conductor-db-sync" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.003145 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.006620 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.010802 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.020817 5035 scope.go:117] "RemoveContainer" containerID="babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05" Nov 24 19:34:47 crc kubenswrapper[5035]: E1124 19:34:47.021227 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05\": container with ID starting with babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05 not found: ID does not exist" containerID="babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.021258 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05"} err="failed to get container status \"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05\": rpc error: code = NotFound desc = could not find container \"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05\": container with ID starting with babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05 not found: ID does not exist" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.021286 5035 scope.go:117] "RemoveContainer" containerID="8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8" Nov 24 19:34:47 crc kubenswrapper[5035]: E1124 19:34:47.025123 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8\": container with ID starting with 8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8 not found: ID does not exist" containerID="8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.025265 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8"} err="failed to get container status \"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8\": rpc error: code = NotFound desc = could not find container \"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8\": container with ID starting with 8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8 not found: ID does not exist" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.025392 5035 scope.go:117] "RemoveContainer" containerID="babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.026206 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05"} err="failed to get container status \"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05\": rpc error: code = NotFound desc = could not find container \"babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05\": container with ID starting with babbd1c52098710cd667be532f66434ac72c47dc87bf22c0e1097b9adb082b05 not found: ID does not exist" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.026254 5035 scope.go:117] "RemoveContainer" containerID="8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.027415 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8"} err="failed to get container status \"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8\": rpc error: code = NotFound desc = could not find container \"8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8\": container with ID starting with 8fd7bf1831cee6d63a2c9b42769988e33b60dd6d97d3d9f63905abedc8ec19c8 not found: ID does not exist" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.033377 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.042160 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.050128 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.051671 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.060118 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.060481 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.061585 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.131390 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjxnj\" (UniqueName: \"kubernetes.io/projected/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-kube-api-access-kjxnj\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.131821 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.131901 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.131942 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-config-data\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.132038 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.132170 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.132213 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8swr\" (UniqueName: \"kubernetes.io/projected/f6fed84b-a0cc-4bae-84af-f318b03d032e-kube-api-access-r8swr\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.132307 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6fed84b-a0cc-4bae-84af-f318b03d032e-logs\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234238 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234357 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8swr\" (UniqueName: \"kubernetes.io/projected/f6fed84b-a0cc-4bae-84af-f318b03d032e-kube-api-access-r8swr\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234439 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6fed84b-a0cc-4bae-84af-f318b03d032e-logs\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234542 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjxnj\" (UniqueName: \"kubernetes.io/projected/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-kube-api-access-kjxnj\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234666 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234823 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234898 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-config-data\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.234953 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.235432 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6fed84b-a0cc-4bae-84af-f318b03d032e-logs\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.238915 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.240327 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-config-data\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.245925 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.248243 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.249653 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.262477 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8swr\" (UniqueName: \"kubernetes.io/projected/f6fed84b-a0cc-4bae-84af-f318b03d032e-kube-api-access-r8swr\") pod \"nova-metadata-0\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.270024 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjxnj\" (UniqueName: \"kubernetes.io/projected/5a4dd86f-4c97-4f2d-b67b-ce7746d105ca-kube-api-access-kjxnj\") pod \"nova-cell1-conductor-0\" (UID: \"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca\") " pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.347554 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.384819 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.828646 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.916794 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.925034 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca","Type":"ContainerStarted","Data":"b29ebb8972ccbc0ede2f5052057b4394121e9f47832164e8a0eec91d9fbff618"} Nov 24 19:34:47 crc kubenswrapper[5035]: I1124 19:34:47.925137 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ed049f4d-b338-46f4-aeb5-853d0bba9e87" containerName="nova-scheduler-scheduler" containerID="cri-o://ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" gracePeriod=30 Nov 24 19:34:47 crc kubenswrapper[5035]: W1124 19:34:47.931488 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6fed84b_a0cc_4bae_84af_f318b03d032e.slice/crio-058ddd22913d1254051827fa9d6388c150e56395d388b09c51e31df10fc0af6e WatchSource:0}: Error finding container 058ddd22913d1254051827fa9d6388c150e56395d388b09c51e31df10fc0af6e: Status 404 returned error can't find the container with id 058ddd22913d1254051827fa9d6388c150e56395d388b09c51e31df10fc0af6e Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.224425 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f00d60b-a125-4768-b123-90859a45671d" path="/var/lib/kubelet/pods/9f00d60b-a125-4768-b123-90859a45671d/volumes" Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.228690 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6cf7f09-7707-4a27-a61e-1d49897c2e63" path="/var/lib/kubelet/pods/c6cf7f09-7707-4a27-a61e-1d49897c2e63/volumes" Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.936804 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6fed84b-a0cc-4bae-84af-f318b03d032e","Type":"ContainerStarted","Data":"759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327"} Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.936850 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6fed84b-a0cc-4bae-84af-f318b03d032e","Type":"ContainerStarted","Data":"b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858"} Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.936865 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6fed84b-a0cc-4bae-84af-f318b03d032e","Type":"ContainerStarted","Data":"058ddd22913d1254051827fa9d6388c150e56395d388b09c51e31df10fc0af6e"} Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.941119 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5a4dd86f-4c97-4f2d-b67b-ce7746d105ca","Type":"ContainerStarted","Data":"31d93602b97d4b9a5531a84080aa2b1f4ce5fcd56af649f6e07082125c4d0d53"} Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.941272 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.974406 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.974384732 podStartE2EDuration="2.974384732s" podCreationTimestamp="2025-11-24 19:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:48.953392052 +0000 UTC m=+1227.475898319" watchObservedRunningTime="2025-11-24 19:34:48.974384732 +0000 UTC m=+1227.496890989" Nov 24 19:34:48 crc kubenswrapper[5035]: I1124 19:34:48.983662 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.983638273 podStartE2EDuration="2.983638273s" podCreationTimestamp="2025-11-24 19:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:48.969971882 +0000 UTC m=+1227.492478139" watchObservedRunningTime="2025-11-24 19:34:48.983638273 +0000 UTC m=+1227.506144530" Nov 24 19:34:50 crc kubenswrapper[5035]: E1124 19:34:50.168213 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 19:34:50 crc kubenswrapper[5035]: E1124 19:34:50.169618 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 19:34:50 crc kubenswrapper[5035]: E1124 19:34:50.170930 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 19:34:50 crc kubenswrapper[5035]: E1124 19:34:50.170995 5035 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ed049f4d-b338-46f4-aeb5-853d0bba9e87" containerName="nova-scheduler-scheduler" Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.942031 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.964422 5035 generic.go:334] "Generic (PLEG): container finished" podID="ed049f4d-b338-46f4-aeb5-853d0bba9e87" containerID="ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" exitCode=0 Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.964461 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed049f4d-b338-46f4-aeb5-853d0bba9e87","Type":"ContainerDied","Data":"ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8"} Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.964476 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.964496 5035 scope.go:117] "RemoveContainer" containerID="ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.964486 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed049f4d-b338-46f4-aeb5-853d0bba9e87","Type":"ContainerDied","Data":"1d7aa5c972260162d39f9909981a9a04af1281e10bf58946080ae2c8e1094f09"} Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.993880 5035 scope.go:117] "RemoveContainer" containerID="ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" Nov 24 19:34:50 crc kubenswrapper[5035]: E1124 19:34:50.994395 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8\": container with ID starting with ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8 not found: ID does not exist" containerID="ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8" Nov 24 19:34:50 crc kubenswrapper[5035]: I1124 19:34:50.994448 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8"} err="failed to get container status \"ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8\": rpc error: code = NotFound desc = could not find container \"ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8\": container with ID starting with ca219f12378578c190ba465b3ac7f0c1db964b03330f973f11dba79b93e1a4b8 not found: ID does not exist" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.021175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qlss\" (UniqueName: \"kubernetes.io/projected/ed049f4d-b338-46f4-aeb5-853d0bba9e87-kube-api-access-4qlss\") pod \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.021247 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-combined-ca-bundle\") pod \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.021431 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-config-data\") pod \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\" (UID: \"ed049f4d-b338-46f4-aeb5-853d0bba9e87\") " Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.039851 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed049f4d-b338-46f4-aeb5-853d0bba9e87-kube-api-access-4qlss" (OuterVolumeSpecName: "kube-api-access-4qlss") pod "ed049f4d-b338-46f4-aeb5-853d0bba9e87" (UID: "ed049f4d-b338-46f4-aeb5-853d0bba9e87"). InnerVolumeSpecName "kube-api-access-4qlss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.066008 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-config-data" (OuterVolumeSpecName: "config-data") pod "ed049f4d-b338-46f4-aeb5-853d0bba9e87" (UID: "ed049f4d-b338-46f4-aeb5-853d0bba9e87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.070538 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed049f4d-b338-46f4-aeb5-853d0bba9e87" (UID: "ed049f4d-b338-46f4-aeb5-853d0bba9e87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.123116 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.123147 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed049f4d-b338-46f4-aeb5-853d0bba9e87-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.123158 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qlss\" (UniqueName: \"kubernetes.io/projected/ed049f4d-b338-46f4-aeb5-853d0bba9e87-kube-api-access-4qlss\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.306045 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.322807 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.344234 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:51 crc kubenswrapper[5035]: E1124 19:34:51.345585 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed049f4d-b338-46f4-aeb5-853d0bba9e87" containerName="nova-scheduler-scheduler" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.345698 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed049f4d-b338-46f4-aeb5-853d0bba9e87" containerName="nova-scheduler-scheduler" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.345999 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed049f4d-b338-46f4-aeb5-853d0bba9e87" containerName="nova-scheduler-scheduler" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.346932 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.349088 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.384980 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.428185 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-config-data\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.428246 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.428268 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwj8z\" (UniqueName: \"kubernetes.io/projected/1a65c0b9-7094-4f6c-85af-6137ccf3d944-kube-api-access-jwj8z\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.530283 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-config-data\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.530607 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.530697 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwj8z\" (UniqueName: \"kubernetes.io/projected/1a65c0b9-7094-4f6c-85af-6137ccf3d944-kube-api-access-jwj8z\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.534755 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.541970 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-config-data\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.551066 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwj8z\" (UniqueName: \"kubernetes.io/projected/1a65c0b9-7094-4f6c-85af-6137ccf3d944-kube-api-access-jwj8z\") pod \"nova-scheduler-0\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " pod="openstack/nova-scheduler-0" Nov 24 19:34:51 crc kubenswrapper[5035]: I1124 19:34:51.674064 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.137540 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.217687 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed049f4d-b338-46f4-aeb5-853d0bba9e87" path="/var/lib/kubelet/pods/ed049f4d-b338-46f4-aeb5-853d0bba9e87/volumes" Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.385081 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.385472 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.981856 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a65c0b9-7094-4f6c-85af-6137ccf3d944","Type":"ContainerStarted","Data":"a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e"} Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.981909 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a65c0b9-7094-4f6c-85af-6137ccf3d944","Type":"ContainerStarted","Data":"0eea632520bb9e31969bb0b3a2637a80895f25541638cefc19720dfc3396197c"} Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.983430 5035 generic.go:334] "Generic (PLEG): container finished" podID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerID="4c0696c2b87e0bd41fbaed214a5bf7b1d0a902951f36ae9b00d9a44f93eb64a9" exitCode=0 Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.983469 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1","Type":"ContainerDied","Data":"4c0696c2b87e0bd41fbaed214a5bf7b1d0a902951f36ae9b00d9a44f93eb64a9"} Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.983493 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1","Type":"ContainerDied","Data":"a6da8996f6b4dec38bd2e3c030b6081d9bd39f0ae9e96bf7a6c4ab8dae5e8e02"} Nov 24 19:34:52 crc kubenswrapper[5035]: I1124 19:34:52.983506 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6da8996f6b4dec38bd2e3c030b6081d9bd39f0ae9e96bf7a6c4ab8dae5e8e02" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.008946 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.008920947 podStartE2EDuration="2.008920947s" podCreationTimestamp="2025-11-24 19:34:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:53.000316583 +0000 UTC m=+1231.522822850" watchObservedRunningTime="2025-11-24 19:34:53.008920947 +0000 UTC m=+1231.531427224" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.078722 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.160808 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-config-data\") pod \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.161606 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-logs\") pod \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.161638 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsz4g\" (UniqueName: \"kubernetes.io/projected/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-kube-api-access-dsz4g\") pod \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.161690 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-combined-ca-bundle\") pod \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\" (UID: \"76892b53-4e77-4bbb-bddd-5f6e98ec2ed1\") " Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.162140 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-logs" (OuterVolumeSpecName: "logs") pod "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" (UID: "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.168958 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-kube-api-access-dsz4g" (OuterVolumeSpecName: "kube-api-access-dsz4g") pod "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" (UID: "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1"). InnerVolumeSpecName "kube-api-access-dsz4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.192668 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.208576 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" (UID: "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.208740 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-config-data" (OuterVolumeSpecName: "config-data") pod "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" (UID: "76892b53-4e77-4bbb-bddd-5f6e98ec2ed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.263470 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.263504 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.263514 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsz4g\" (UniqueName: \"kubernetes.io/projected/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-kube-api-access-dsz4g\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.263524 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:34:53 crc kubenswrapper[5035]: I1124 19:34:53.994702 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.036838 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.047986 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.062435 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:54 crc kubenswrapper[5035]: E1124 19:34:54.063344 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-log" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.063490 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-log" Nov 24 19:34:54 crc kubenswrapper[5035]: E1124 19:34:54.063566 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-api" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.063575 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-api" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.063836 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-api" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.063862 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" containerName="nova-api-log" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.065259 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.067942 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.080720 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.182237 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ca8fc75-38aa-4491-879d-6b41a4d440fd-logs\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.182365 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdmdg\" (UniqueName: \"kubernetes.io/projected/0ca8fc75-38aa-4491-879d-6b41a4d440fd-kube-api-access-pdmdg\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.182452 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.182481 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-config-data\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.212107 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76892b53-4e77-4bbb-bddd-5f6e98ec2ed1" path="/var/lib/kubelet/pods/76892b53-4e77-4bbb-bddd-5f6e98ec2ed1/volumes" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.284172 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ca8fc75-38aa-4491-879d-6b41a4d440fd-logs\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.284655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdmdg\" (UniqueName: \"kubernetes.io/projected/0ca8fc75-38aa-4491-879d-6b41a4d440fd-kube-api-access-pdmdg\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.284722 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ca8fc75-38aa-4491-879d-6b41a4d440fd-logs\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.284778 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.284814 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-config-data\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.289951 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-config-data\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.303030 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.319229 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdmdg\" (UniqueName: \"kubernetes.io/projected/0ca8fc75-38aa-4491-879d-6b41a4d440fd-kube-api-access-pdmdg\") pod \"nova-api-0\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.389158 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:34:54 crc kubenswrapper[5035]: W1124 19:34:54.953807 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ca8fc75_38aa_4491_879d_6b41a4d440fd.slice/crio-8adf14fc3fad3d87f3d8e34e104162d5a1d1c115c95ddbf22073e7a249c44d9c WatchSource:0}: Error finding container 8adf14fc3fad3d87f3d8e34e104162d5a1d1c115c95ddbf22073e7a249c44d9c: Status 404 returned error can't find the container with id 8adf14fc3fad3d87f3d8e34e104162d5a1d1c115c95ddbf22073e7a249c44d9c Nov 24 19:34:54 crc kubenswrapper[5035]: I1124 19:34:54.956146 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:34:55 crc kubenswrapper[5035]: I1124 19:34:55.005980 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ca8fc75-38aa-4491-879d-6b41a4d440fd","Type":"ContainerStarted","Data":"8adf14fc3fad3d87f3d8e34e104162d5a1d1c115c95ddbf22073e7a249c44d9c"} Nov 24 19:34:56 crc kubenswrapper[5035]: I1124 19:34:56.019754 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ca8fc75-38aa-4491-879d-6b41a4d440fd","Type":"ContainerStarted","Data":"0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0"} Nov 24 19:34:56 crc kubenswrapper[5035]: I1124 19:34:56.020426 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ca8fc75-38aa-4491-879d-6b41a4d440fd","Type":"ContainerStarted","Data":"f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77"} Nov 24 19:34:56 crc kubenswrapper[5035]: I1124 19:34:56.050951 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.050918269 podStartE2EDuration="2.050918269s" podCreationTimestamp="2025-11-24 19:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:34:56.038792321 +0000 UTC m=+1234.561298588" watchObservedRunningTime="2025-11-24 19:34:56.050918269 +0000 UTC m=+1234.573424576" Nov 24 19:34:56 crc kubenswrapper[5035]: I1124 19:34:56.675103 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 19:34:57 crc kubenswrapper[5035]: I1124 19:34:57.385673 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 19:34:57 crc kubenswrapper[5035]: I1124 19:34:57.386164 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 19:34:57 crc kubenswrapper[5035]: I1124 19:34:57.386238 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 19:34:58 crc kubenswrapper[5035]: I1124 19:34:58.406485 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 19:34:58 crc kubenswrapper[5035]: I1124 19:34:58.406544 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 19:35:01 crc kubenswrapper[5035]: I1124 19:35:01.675199 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 19:35:01 crc kubenswrapper[5035]: I1124 19:35:01.713940 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 19:35:02 crc kubenswrapper[5035]: I1124 19:35:02.142656 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 19:35:04 crc kubenswrapper[5035]: I1124 19:35:04.390103 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 19:35:04 crc kubenswrapper[5035]: I1124 19:35:04.390481 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 19:35:05 crc kubenswrapper[5035]: I1124 19:35:05.472481 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 19:35:05 crc kubenswrapper[5035]: I1124 19:35:05.472809 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 19:35:07 crc kubenswrapper[5035]: I1124 19:35:07.392036 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 19:35:07 crc kubenswrapper[5035]: I1124 19:35:07.396139 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 19:35:07 crc kubenswrapper[5035]: I1124 19:35:07.399206 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 19:35:08 crc kubenswrapper[5035]: I1124 19:35:08.155585 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.142220 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.170014 5035 generic.go:334] "Generic (PLEG): container finished" podID="30ab1877-0558-4e64-8ea3-157b2d699faa" containerID="d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4" exitCode=137 Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.170083 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.170086 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30ab1877-0558-4e64-8ea3-157b2d699faa","Type":"ContainerDied","Data":"d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4"} Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.170158 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"30ab1877-0558-4e64-8ea3-157b2d699faa","Type":"ContainerDied","Data":"fe93ab15c638376edb37ed53b9ceff577ad141b96d503e5481fa5eec35e32346"} Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.170188 5035 scope.go:117] "RemoveContainer" containerID="d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.207125 5035 scope.go:117] "RemoveContainer" containerID="d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4" Nov 24 19:35:10 crc kubenswrapper[5035]: E1124 19:35:10.207736 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4\": container with ID starting with d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4 not found: ID does not exist" containerID="d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.207790 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4"} err="failed to get container status \"d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4\": rpc error: code = NotFound desc = could not find container \"d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4\": container with ID starting with d8d7e759868c44417d2ce07105b34212a9f05d5f943d2faf484d108b2e3134b4 not found: ID does not exist" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.288673 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-config-data\") pod \"30ab1877-0558-4e64-8ea3-157b2d699faa\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.288814 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdcsn\" (UniqueName: \"kubernetes.io/projected/30ab1877-0558-4e64-8ea3-157b2d699faa-kube-api-access-vdcsn\") pod \"30ab1877-0558-4e64-8ea3-157b2d699faa\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.288872 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-combined-ca-bundle\") pod \"30ab1877-0558-4e64-8ea3-157b2d699faa\" (UID: \"30ab1877-0558-4e64-8ea3-157b2d699faa\") " Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.295090 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30ab1877-0558-4e64-8ea3-157b2d699faa-kube-api-access-vdcsn" (OuterVolumeSpecName: "kube-api-access-vdcsn") pod "30ab1877-0558-4e64-8ea3-157b2d699faa" (UID: "30ab1877-0558-4e64-8ea3-157b2d699faa"). InnerVolumeSpecName "kube-api-access-vdcsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.315754 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-config-data" (OuterVolumeSpecName: "config-data") pod "30ab1877-0558-4e64-8ea3-157b2d699faa" (UID: "30ab1877-0558-4e64-8ea3-157b2d699faa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.320067 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30ab1877-0558-4e64-8ea3-157b2d699faa" (UID: "30ab1877-0558-4e64-8ea3-157b2d699faa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.393450 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdcsn\" (UniqueName: \"kubernetes.io/projected/30ab1877-0558-4e64-8ea3-157b2d699faa-kube-api-access-vdcsn\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.393513 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.393532 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30ab1877-0558-4e64-8ea3-157b2d699faa-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.515627 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.526501 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.540423 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:35:10 crc kubenswrapper[5035]: E1124 19:35:10.540865 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30ab1877-0558-4e64-8ea3-157b2d699faa" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.540885 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="30ab1877-0558-4e64-8ea3-157b2d699faa" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.541052 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="30ab1877-0558-4e64-8ea3-157b2d699faa" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.541660 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.545413 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.548426 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.549687 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.549839 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.698468 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.698513 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.698545 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.698683 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.698921 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27xwm\" (UniqueName: \"kubernetes.io/projected/6a38d999-4d5c-47ad-91af-605b01a6daab-kube-api-access-27xwm\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.800467 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.801332 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.801412 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.801447 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.801495 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27xwm\" (UniqueName: \"kubernetes.io/projected/6a38d999-4d5c-47ad-91af-605b01a6daab-kube-api-access-27xwm\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.805004 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.805334 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.805747 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.805768 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38d999-4d5c-47ad-91af-605b01a6daab-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.826121 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27xwm\" (UniqueName: \"kubernetes.io/projected/6a38d999-4d5c-47ad-91af-605b01a6daab-kube-api-access-27xwm\") pod \"nova-cell1-novncproxy-0\" (UID: \"6a38d999-4d5c-47ad-91af-605b01a6daab\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:10 crc kubenswrapper[5035]: I1124 19:35:10.855981 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:11 crc kubenswrapper[5035]: I1124 19:35:11.320630 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 19:35:12 crc kubenswrapper[5035]: I1124 19:35:12.191651 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6a38d999-4d5c-47ad-91af-605b01a6daab","Type":"ContainerStarted","Data":"73720effce7c6a05de53c9b51a6d5a0db594a69961a269ef85ce59ac8c998d00"} Nov 24 19:35:12 crc kubenswrapper[5035]: I1124 19:35:12.192104 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6a38d999-4d5c-47ad-91af-605b01a6daab","Type":"ContainerStarted","Data":"b4c0e3a3af382489f80f704b4df0789b457a4027242aac53c4a3a55a2cc56cf6"} Nov 24 19:35:12 crc kubenswrapper[5035]: I1124 19:35:12.212540 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.212521708 podStartE2EDuration="2.212521708s" podCreationTimestamp="2025-11-24 19:35:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:35:12.208313004 +0000 UTC m=+1250.730819251" watchObservedRunningTime="2025-11-24 19:35:12.212521708 +0000 UTC m=+1250.735027985" Nov 24 19:35:12 crc kubenswrapper[5035]: I1124 19:35:12.221934 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30ab1877-0558-4e64-8ea3-157b2d699faa" path="/var/lib/kubelet/pods/30ab1877-0558-4e64-8ea3-157b2d699faa/volumes" Nov 24 19:35:14 crc kubenswrapper[5035]: I1124 19:35:14.394413 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 19:35:14 crc kubenswrapper[5035]: I1124 19:35:14.394991 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 19:35:14 crc kubenswrapper[5035]: I1124 19:35:14.396267 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 19:35:14 crc kubenswrapper[5035]: I1124 19:35:14.397501 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.214564 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.218740 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.400542 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f95c456cf-dz6jb"] Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.426673 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.428214 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f95c456cf-dz6jb"] Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.513433 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg7wd\" (UniqueName: \"kubernetes.io/projected/0e10dfb5-71c9-4b4c-a520-0845163d3b42-kube-api-access-cg7wd\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.513541 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-nb\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.513559 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-dns-svc\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.513800 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-sb\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.513983 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-config\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.615931 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg7wd\" (UniqueName: \"kubernetes.io/projected/0e10dfb5-71c9-4b4c-a520-0845163d3b42-kube-api-access-cg7wd\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.616070 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-nb\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.616095 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-dns-svc\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.616148 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-sb\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.616193 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-config\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.618182 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-config\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.619092 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-nb\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.619709 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-dns-svc\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.620573 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-sb\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.639044 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg7wd\" (UniqueName: \"kubernetes.io/projected/0e10dfb5-71c9-4b4c-a520-0845163d3b42-kube-api-access-cg7wd\") pod \"dnsmasq-dns-f95c456cf-dz6jb\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.752596 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:15 crc kubenswrapper[5035]: I1124 19:35:15.857076 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:16 crc kubenswrapper[5035]: I1124 19:35:16.263495 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f95c456cf-dz6jb"] Nov 24 19:35:16 crc kubenswrapper[5035]: I1124 19:35:16.755174 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:16 crc kubenswrapper[5035]: I1124 19:35:16.755844 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-central-agent" containerID="cri-o://4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397" gracePeriod=30 Nov 24 19:35:16 crc kubenswrapper[5035]: I1124 19:35:16.755997 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="proxy-httpd" containerID="cri-o://b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e" gracePeriod=30 Nov 24 19:35:16 crc kubenswrapper[5035]: I1124 19:35:16.756051 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="sg-core" containerID="cri-o://8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53" gracePeriod=30 Nov 24 19:35:16 crc kubenswrapper[5035]: I1124 19:35:16.756095 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-notification-agent" containerID="cri-o://1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab" gracePeriod=30 Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.233260 5035 generic.go:334] "Generic (PLEG): container finished" podID="2554287e-1a71-419a-95c0-260b69535239" containerID="b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e" exitCode=0 Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.233332 5035 generic.go:334] "Generic (PLEG): container finished" podID="2554287e-1a71-419a-95c0-260b69535239" containerID="8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53" exitCode=2 Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.233340 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerDied","Data":"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e"} Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.233388 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerDied","Data":"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53"} Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.234881 5035 generic.go:334] "Generic (PLEG): container finished" podID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerID="2b5e3d5b19b7e67599e8bc8b7f2ccc8d2971262f25bbf5955762dbb2425679d9" exitCode=0 Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.234918 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" event={"ID":"0e10dfb5-71c9-4b4c-a520-0845163d3b42","Type":"ContainerDied","Data":"2b5e3d5b19b7e67599e8bc8b7f2ccc8d2971262f25bbf5955762dbb2425679d9"} Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.235035 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" event={"ID":"0e10dfb5-71c9-4b4c-a520-0845163d3b42","Type":"ContainerStarted","Data":"7dc2a70ad99427b333b728505aa691b1d495c0b5fb1454e0015074b48b888652"} Nov 24 19:35:17 crc kubenswrapper[5035]: I1124 19:35:17.836843 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.145307 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.158229 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-log-httpd\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.158407 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4crfh\" (UniqueName: \"kubernetes.io/projected/2554287e-1a71-419a-95c0-260b69535239-kube-api-access-4crfh\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.158536 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-ceilometer-tls-certs\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.158578 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-sg-core-conf-yaml\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.158894 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.163063 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-config-data\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.163112 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-scripts\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.163158 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-run-httpd\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.163215 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-combined-ca-bundle\") pod \"2554287e-1a71-419a-95c0-260b69535239\" (UID: \"2554287e-1a71-419a-95c0-260b69535239\") " Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.164110 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.164156 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.164261 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2554287e-1a71-419a-95c0-260b69535239-kube-api-access-4crfh" (OuterVolumeSpecName: "kube-api-access-4crfh") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "kube-api-access-4crfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.167474 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-scripts" (OuterVolumeSpecName: "scripts") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.208383 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.248593 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.254231 5035 generic.go:334] "Generic (PLEG): container finished" podID="2554287e-1a71-419a-95c0-260b69535239" containerID="1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab" exitCode=0 Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.254323 5035 generic.go:334] "Generic (PLEG): container finished" podID="2554287e-1a71-419a-95c0-260b69535239" containerID="4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397" exitCode=0 Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.254464 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.254726 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerDied","Data":"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab"} Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.254755 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerDied","Data":"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397"} Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.254765 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2554287e-1a71-419a-95c0-260b69535239","Type":"ContainerDied","Data":"e33563dfd2e612766741e01f83337055e1d71bc85450b8494badd9d4060b7879"} Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.254778 5035 scope.go:117] "RemoveContainer" containerID="b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.259259 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-log" containerID="cri-o://f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77" gracePeriod=30 Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.259548 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" event={"ID":"0e10dfb5-71c9-4b4c-a520-0845163d3b42","Type":"ContainerStarted","Data":"93b09cd18ef0bac4192a0c4d0ae1a671fd2cfc5fd437e2653f4e0176ad86ee34"} Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.259711 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-api" containerID="cri-o://0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0" gracePeriod=30 Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.260217 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.268734 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4crfh\" (UniqueName: \"kubernetes.io/projected/2554287e-1a71-419a-95c0-260b69535239-kube-api-access-4crfh\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.268785 5035 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.268801 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.268812 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.268825 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2554287e-1a71-419a-95c0-260b69535239-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.295536 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" podStartSLOduration=3.295513785 podStartE2EDuration="3.295513785s" podCreationTimestamp="2025-11-24 19:35:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:35:18.292843772 +0000 UTC m=+1256.815350039" watchObservedRunningTime="2025-11-24 19:35:18.295513785 +0000 UTC m=+1256.818020042" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.298211 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-config-data" (OuterVolumeSpecName: "config-data") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.298273 5035 scope.go:117] "RemoveContainer" containerID="8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.320498 5035 scope.go:117] "RemoveContainer" containerID="1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.324393 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2554287e-1a71-419a-95c0-260b69535239" (UID: "2554287e-1a71-419a-95c0-260b69535239"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.339606 5035 scope.go:117] "RemoveContainer" containerID="4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.361736 5035 scope.go:117] "RemoveContainer" containerID="b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e" Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.362201 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e\": container with ID starting with b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e not found: ID does not exist" containerID="b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.362235 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e"} err="failed to get container status \"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e\": rpc error: code = NotFound desc = could not find container \"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e\": container with ID starting with b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.362254 5035 scope.go:117] "RemoveContainer" containerID="8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53" Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.362487 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53\": container with ID starting with 8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53 not found: ID does not exist" containerID="8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.362509 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53"} err="failed to get container status \"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53\": rpc error: code = NotFound desc = could not find container \"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53\": container with ID starting with 8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53 not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.362522 5035 scope.go:117] "RemoveContainer" containerID="1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab" Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.362904 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab\": container with ID starting with 1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab not found: ID does not exist" containerID="1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.362931 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab"} err="failed to get container status \"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab\": rpc error: code = NotFound desc = could not find container \"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab\": container with ID starting with 1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.362949 5035 scope.go:117] "RemoveContainer" containerID="4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397" Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.363243 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397\": container with ID starting with 4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397 not found: ID does not exist" containerID="4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.363366 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397"} err="failed to get container status \"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397\": rpc error: code = NotFound desc = could not find container \"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397\": container with ID starting with 4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397 not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.363458 5035 scope.go:117] "RemoveContainer" containerID="b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.363860 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e"} err="failed to get container status \"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e\": rpc error: code = NotFound desc = could not find container \"b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e\": container with ID starting with b82e6429bf68d3187f06416e5286a42ea25125e7b0daba84859947a78d9c574e not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.363883 5035 scope.go:117] "RemoveContainer" containerID="8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.364098 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53"} err="failed to get container status \"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53\": rpc error: code = NotFound desc = could not find container \"8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53\": container with ID starting with 8de098251bf62bc8889742808fe436252f77e1860c638341cec2e386fc5abd53 not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.364195 5035 scope.go:117] "RemoveContainer" containerID="1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.364628 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab"} err="failed to get container status \"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab\": rpc error: code = NotFound desc = could not find container \"1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab\": container with ID starting with 1a15d0f68cb01da9769a818aed330ae2185919a5f2fab929f8e077245a2996ab not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.364647 5035 scope.go:117] "RemoveContainer" containerID="4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.364904 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397"} err="failed to get container status \"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397\": rpc error: code = NotFound desc = could not find container \"4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397\": container with ID starting with 4ecfbb7a1e53aabf9241e71f85b614f19b2a298a4df029f3b8b0da0299352397 not found: ID does not exist" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.371124 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.371164 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2554287e-1a71-419a-95c0-260b69535239-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.588144 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.594786 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.611327 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.611841 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-central-agent" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.611858 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-central-agent" Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.611869 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-notification-agent" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.611876 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-notification-agent" Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.611893 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="proxy-httpd" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.611899 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="proxy-httpd" Nov 24 19:35:18 crc kubenswrapper[5035]: E1124 19:35:18.611920 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="sg-core" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.611925 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="sg-core" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.612073 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="proxy-httpd" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.612086 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-notification-agent" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.612097 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="ceilometer-central-agent" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.612109 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2554287e-1a71-419a-95c0-260b69535239" containerName="sg-core" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.613739 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.615623 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.615764 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.616901 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.627330 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685001 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvpg8\" (UniqueName: \"kubernetes.io/projected/286dd62a-ea0d-4d35-ac51-ff814de3eef0-kube-api-access-mvpg8\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685272 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-scripts\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685322 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685343 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685394 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-config-data\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685423 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-run-httpd\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685446 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-log-httpd\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.685461 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787453 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-config-data\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787518 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-run-httpd\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787547 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-log-httpd\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787562 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787622 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvpg8\" (UniqueName: \"kubernetes.io/projected/286dd62a-ea0d-4d35-ac51-ff814de3eef0-kube-api-access-mvpg8\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787647 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-scripts\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787675 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.787709 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.788271 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-log-httpd\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.788266 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-run-httpd\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.791537 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.791546 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-scripts\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.791614 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.796857 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.804712 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-config-data\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:18 crc kubenswrapper[5035]: I1124 19:35:18.807851 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvpg8\" (UniqueName: \"kubernetes.io/projected/286dd62a-ea0d-4d35-ac51-ff814de3eef0-kube-api-access-mvpg8\") pod \"ceilometer-0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " pod="openstack/ceilometer-0" Nov 24 19:35:19 crc kubenswrapper[5035]: I1124 19:35:19.023872 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:19 crc kubenswrapper[5035]: I1124 19:35:19.024585 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:19 crc kubenswrapper[5035]: I1124 19:35:19.275129 5035 generic.go:334] "Generic (PLEG): container finished" podID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerID="f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77" exitCode=143 Nov 24 19:35:19 crc kubenswrapper[5035]: I1124 19:35:19.275496 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ca8fc75-38aa-4491-879d-6b41a4d440fd","Type":"ContainerDied","Data":"f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77"} Nov 24 19:35:19 crc kubenswrapper[5035]: I1124 19:35:19.495227 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:20 crc kubenswrapper[5035]: I1124 19:35:20.211284 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2554287e-1a71-419a-95c0-260b69535239" path="/var/lib/kubelet/pods/2554287e-1a71-419a-95c0-260b69535239/volumes" Nov 24 19:35:20 crc kubenswrapper[5035]: I1124 19:35:20.286304 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerStarted","Data":"4856480726a423da01c3655a262332b2078a84b354378cce3495899f7c2960a4"} Nov 24 19:35:20 crc kubenswrapper[5035]: I1124 19:35:20.286357 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerStarted","Data":"0bfd7a726db1b7deaaff98ba31ce4ebd244bbff88512c3f1b85ddc7f9f31c666"} Nov 24 19:35:20 crc kubenswrapper[5035]: I1124 19:35:20.856326 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:20 crc kubenswrapper[5035]: I1124 19:35:20.892796 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.298055 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerStarted","Data":"79b3c0a93f919b8ed4792288c2a4d902a251d8e956acc16231886fbf335eef7f"} Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.317421 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.533162 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-gnkh9"] Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.534683 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.538383 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.538495 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.545555 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gnkh9"] Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.637428 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.637740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-config-data\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.637911 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-scripts\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.638197 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79jrn\" (UniqueName: \"kubernetes.io/projected/99e36d14-3aed-4554-8cab-f20fa21a29d8-kube-api-access-79jrn\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.739544 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79jrn\" (UniqueName: \"kubernetes.io/projected/99e36d14-3aed-4554-8cab-f20fa21a29d8-kube-api-access-79jrn\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.739626 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.739653 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-config-data\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.739700 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-scripts\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.748355 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-scripts\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.754134 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.754600 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-config-data\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.773962 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79jrn\" (UniqueName: \"kubernetes.io/projected/99e36d14-3aed-4554-8cab-f20fa21a29d8-kube-api-access-79jrn\") pod \"nova-cell1-cell-mapping-gnkh9\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:21 crc kubenswrapper[5035]: E1124 19:35:21.841686 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ca8fc75_38aa_4491_879d_6b41a4d440fd.slice/crio-conmon-0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0.scope\": RecentStats: unable to find data in memory cache]" Nov 24 19:35:21 crc kubenswrapper[5035]: I1124 19:35:21.915812 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.046324 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.146876 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-config-data\") pod \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.147308 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdmdg\" (UniqueName: \"kubernetes.io/projected/0ca8fc75-38aa-4491-879d-6b41a4d440fd-kube-api-access-pdmdg\") pod \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.147457 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ca8fc75-38aa-4491-879d-6b41a4d440fd-logs\") pod \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.147485 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-combined-ca-bundle\") pod \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\" (UID: \"0ca8fc75-38aa-4491-879d-6b41a4d440fd\") " Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.149053 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca8fc75-38aa-4491-879d-6b41a4d440fd-logs" (OuterVolumeSpecName: "logs") pod "0ca8fc75-38aa-4491-879d-6b41a4d440fd" (UID: "0ca8fc75-38aa-4491-879d-6b41a4d440fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.178646 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ca8fc75-38aa-4491-879d-6b41a4d440fd" (UID: "0ca8fc75-38aa-4491-879d-6b41a4d440fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.187585 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca8fc75-38aa-4491-879d-6b41a4d440fd-kube-api-access-pdmdg" (OuterVolumeSpecName: "kube-api-access-pdmdg") pod "0ca8fc75-38aa-4491-879d-6b41a4d440fd" (UID: "0ca8fc75-38aa-4491-879d-6b41a4d440fd"). InnerVolumeSpecName "kube-api-access-pdmdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.191402 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-config-data" (OuterVolumeSpecName: "config-data") pod "0ca8fc75-38aa-4491-879d-6b41a4d440fd" (UID: "0ca8fc75-38aa-4491-879d-6b41a4d440fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.249740 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ca8fc75-38aa-4491-879d-6b41a4d440fd-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.249780 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.249795 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca8fc75-38aa-4491-879d-6b41a4d440fd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.249809 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdmdg\" (UniqueName: \"kubernetes.io/projected/0ca8fc75-38aa-4491-879d-6b41a4d440fd-kube-api-access-pdmdg\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.313054 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerStarted","Data":"0507d5c07ce9fb02a2b394735925f16b0843a9d3b866d91373902c03648c34e1"} Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.315251 5035 generic.go:334] "Generic (PLEG): container finished" podID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerID="0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0" exitCode=0 Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.316262 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.316880 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ca8fc75-38aa-4491-879d-6b41a4d440fd","Type":"ContainerDied","Data":"0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0"} Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.316914 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ca8fc75-38aa-4491-879d-6b41a4d440fd","Type":"ContainerDied","Data":"8adf14fc3fad3d87f3d8e34e104162d5a1d1c115c95ddbf22073e7a249c44d9c"} Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.316934 5035 scope.go:117] "RemoveContainer" containerID="0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.342701 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.349253 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.372579 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:22 crc kubenswrapper[5035]: E1124 19:35:22.372979 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-log" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.372991 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-log" Nov 24 19:35:22 crc kubenswrapper[5035]: E1124 19:35:22.373004 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-api" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.373010 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-api" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.373168 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-log" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.373181 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" containerName="nova-api-api" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.374415 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.376284 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.376500 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.378739 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.400182 5035 scope.go:117] "RemoveContainer" containerID="f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.425134 5035 scope.go:117] "RemoveContainer" containerID="0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.425362 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:22 crc kubenswrapper[5035]: E1124 19:35:22.428399 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0\": container with ID starting with 0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0 not found: ID does not exist" containerID="0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.428439 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0"} err="failed to get container status \"0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0\": rpc error: code = NotFound desc = could not find container \"0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0\": container with ID starting with 0e04e3af87d63c04febeb12c2f36e66f78635aebdd98ba5e9ef53895a4f113d0 not found: ID does not exist" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.428461 5035 scope.go:117] "RemoveContainer" containerID="f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77" Nov 24 19:35:22 crc kubenswrapper[5035]: E1124 19:35:22.431457 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77\": container with ID starting with f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77 not found: ID does not exist" containerID="f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.431511 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77"} err="failed to get container status \"f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77\": rpc error: code = NotFound desc = could not find container \"f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77\": container with ID starting with f78c324fe1fc970df393e16e24cb75cf2c3c653673af5d165829e44c29024a77 not found: ID does not exist" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.453605 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-logs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.453651 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-public-tls-certs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.453785 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.453854 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-config-data\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.453924 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nw6v\" (UniqueName: \"kubernetes.io/projected/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-kube-api-access-8nw6v\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.453952 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.466490 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gnkh9"] Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.555761 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-config-data\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.555814 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nw6v\" (UniqueName: \"kubernetes.io/projected/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-kube-api-access-8nw6v\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.555841 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.555889 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-logs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.555908 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-public-tls-certs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.556021 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.556515 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-logs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.559913 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.562129 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-config-data\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.562724 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-public-tls-certs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.563693 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.573761 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nw6v\" (UniqueName: \"kubernetes.io/projected/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-kube-api-access-8nw6v\") pod \"nova-api-0\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " pod="openstack/nova-api-0" Nov 24 19:35:22 crc kubenswrapper[5035]: I1124 19:35:22.726358 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:23 crc kubenswrapper[5035]: I1124 19:35:23.212998 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:23 crc kubenswrapper[5035]: I1124 19:35:23.324350 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f","Type":"ContainerStarted","Data":"a239b0fda47def74fc0c7e6693b7d5ec0359812de2c1338baaaaa1ca19692cb9"} Nov 24 19:35:23 crc kubenswrapper[5035]: I1124 19:35:23.325819 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gnkh9" event={"ID":"99e36d14-3aed-4554-8cab-f20fa21a29d8","Type":"ContainerStarted","Data":"b2a4c21951ea187c9e40b0b6d4b3ce1c29e72c640188b0c8eeed8acf139c21ce"} Nov 24 19:35:23 crc kubenswrapper[5035]: I1124 19:35:23.325850 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gnkh9" event={"ID":"99e36d14-3aed-4554-8cab-f20fa21a29d8","Type":"ContainerStarted","Data":"98a21a092692ecdd8e983256d4e18629ca6b622ad53cdecfd2b7efa418ad0b1e"} Nov 24 19:35:23 crc kubenswrapper[5035]: I1124 19:35:23.350853 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-gnkh9" podStartSLOduration=2.350838643 podStartE2EDuration="2.350838643s" podCreationTimestamp="2025-11-24 19:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:35:23.344848761 +0000 UTC m=+1261.867355018" watchObservedRunningTime="2025-11-24 19:35:23.350838643 +0000 UTC m=+1261.873344900" Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.212663 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca8fc75-38aa-4491-879d-6b41a4d440fd" path="/var/lib/kubelet/pods/0ca8fc75-38aa-4491-879d-6b41a4d440fd/volumes" Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.344159 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f","Type":"ContainerStarted","Data":"059cc1ecdc7b970cadee29e2db971d102d536c8312737643c621bc251aae3f34"} Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.344225 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f","Type":"ContainerStarted","Data":"160a7bfaec1325b20eb24874f945f585a3af874dfd87562a5b8164bdd9d6d7e0"} Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.351570 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-central-agent" containerID="cri-o://4856480726a423da01c3655a262332b2078a84b354378cce3495899f7c2960a4" gracePeriod=30 Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.352196 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerStarted","Data":"7acc11f1e630c81f4a011df3c31e54e0128d9c6eea8c006ab4f842c1d719077f"} Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.352332 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.352437 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="proxy-httpd" containerID="cri-o://7acc11f1e630c81f4a011df3c31e54e0128d9c6eea8c006ab4f842c1d719077f" gracePeriod=30 Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.352560 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="sg-core" containerID="cri-o://0507d5c07ce9fb02a2b394735925f16b0843a9d3b866d91373902c03648c34e1" gracePeriod=30 Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.352655 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-notification-agent" containerID="cri-o://79b3c0a93f919b8ed4792288c2a4d902a251d8e956acc16231886fbf335eef7f" gracePeriod=30 Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.373237 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.373217518 podStartE2EDuration="2.373217518s" podCreationTimestamp="2025-11-24 19:35:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:35:24.365981421 +0000 UTC m=+1262.888487678" watchObservedRunningTime="2025-11-24 19:35:24.373217518 +0000 UTC m=+1262.895723776" Nov 24 19:35:24 crc kubenswrapper[5035]: I1124 19:35:24.394392 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.274293555 podStartE2EDuration="6.394364722s" podCreationTimestamp="2025-11-24 19:35:18 +0000 UTC" firstStartedPulling="2025-11-24 19:35:19.499014773 +0000 UTC m=+1258.021521030" lastFinishedPulling="2025-11-24 19:35:23.61908595 +0000 UTC m=+1262.141592197" observedRunningTime="2025-11-24 19:35:24.389407067 +0000 UTC m=+1262.911913334" watchObservedRunningTime="2025-11-24 19:35:24.394364722 +0000 UTC m=+1262.916870999" Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.362658 5035 generic.go:334] "Generic (PLEG): container finished" podID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerID="7acc11f1e630c81f4a011df3c31e54e0128d9c6eea8c006ab4f842c1d719077f" exitCode=0 Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.362933 5035 generic.go:334] "Generic (PLEG): container finished" podID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerID="0507d5c07ce9fb02a2b394735925f16b0843a9d3b866d91373902c03648c34e1" exitCode=2 Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.362943 5035 generic.go:334] "Generic (PLEG): container finished" podID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerID="79b3c0a93f919b8ed4792288c2a4d902a251d8e956acc16231886fbf335eef7f" exitCode=0 Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.362737 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerDied","Data":"7acc11f1e630c81f4a011df3c31e54e0128d9c6eea8c006ab4f842c1d719077f"} Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.363021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerDied","Data":"0507d5c07ce9fb02a2b394735925f16b0843a9d3b866d91373902c03648c34e1"} Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.363046 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerDied","Data":"79b3c0a93f919b8ed4792288c2a4d902a251d8e956acc16231886fbf335eef7f"} Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.755542 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.835873 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f7bbc55bc-tf5sn"] Nov 24 19:35:25 crc kubenswrapper[5035]: I1124 19:35:25.837203 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" podUID="301002c3-71f8-41bc-bf96-babd19cc994c" containerName="dnsmasq-dns" containerID="cri-o://8fbb254c0db5b18f0d54fed5e8b6cf87978057bd5cd5afc94c18688ea509368a" gracePeriod=10 Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.376907 5035 generic.go:334] "Generic (PLEG): container finished" podID="301002c3-71f8-41bc-bf96-babd19cc994c" containerID="8fbb254c0db5b18f0d54fed5e8b6cf87978057bd5cd5afc94c18688ea509368a" exitCode=0 Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.377305 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" event={"ID":"301002c3-71f8-41bc-bf96-babd19cc994c","Type":"ContainerDied","Data":"8fbb254c0db5b18f0d54fed5e8b6cf87978057bd5cd5afc94c18688ea509368a"} Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.377331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" event={"ID":"301002c3-71f8-41bc-bf96-babd19cc994c","Type":"ContainerDied","Data":"25eee20c6e6bd20ed7a0d8b07bb828834923fe78b1e11ca3b82b034396d637f4"} Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.377343 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25eee20c6e6bd20ed7a0d8b07bb828834923fe78b1e11ca3b82b034396d637f4" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.383826 5035 generic.go:334] "Generic (PLEG): container finished" podID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerID="4856480726a423da01c3655a262332b2078a84b354378cce3495899f7c2960a4" exitCode=0 Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.383864 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerDied","Data":"4856480726a423da01c3655a262332b2078a84b354378cce3495899f7c2960a4"} Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.383889 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"286dd62a-ea0d-4d35-ac51-ff814de3eef0","Type":"ContainerDied","Data":"0bfd7a726db1b7deaaff98ba31ce4ebd244bbff88512c3f1b85ddc7f9f31c666"} Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.383899 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bfd7a726db1b7deaaff98ba31ce4ebd244bbff88512c3f1b85ddc7f9f31c666" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.401913 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.408714 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.527148 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-dns-svc\") pod \"301002c3-71f8-41bc-bf96-babd19cc994c\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.527476 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-nb\") pod \"301002c3-71f8-41bc-bf96-babd19cc994c\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.527521 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-scripts\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.527598 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-combined-ca-bundle\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.530616 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkdbb\" (UniqueName: \"kubernetes.io/projected/301002c3-71f8-41bc-bf96-babd19cc994c-kube-api-access-mkdbb\") pod \"301002c3-71f8-41bc-bf96-babd19cc994c\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531119 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-sb\") pod \"301002c3-71f8-41bc-bf96-babd19cc994c\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531177 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-ceilometer-tls-certs\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531212 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-sg-core-conf-yaml\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531320 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-config\") pod \"301002c3-71f8-41bc-bf96-babd19cc994c\" (UID: \"301002c3-71f8-41bc-bf96-babd19cc994c\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531360 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-run-httpd\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531422 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-config-data\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531453 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-log-httpd\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.531502 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvpg8\" (UniqueName: \"kubernetes.io/projected/286dd62a-ea0d-4d35-ac51-ff814de3eef0-kube-api-access-mvpg8\") pod \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\" (UID: \"286dd62a-ea0d-4d35-ac51-ff814de3eef0\") " Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.532841 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.532947 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-scripts" (OuterVolumeSpecName: "scripts") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.533550 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.533844 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/301002c3-71f8-41bc-bf96-babd19cc994c-kube-api-access-mkdbb" (OuterVolumeSpecName: "kube-api-access-mkdbb") pod "301002c3-71f8-41bc-bf96-babd19cc994c" (UID: "301002c3-71f8-41bc-bf96-babd19cc994c"). InnerVolumeSpecName "kube-api-access-mkdbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.541189 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286dd62a-ea0d-4d35-ac51-ff814de3eef0-kube-api-access-mvpg8" (OuterVolumeSpecName: "kube-api-access-mvpg8") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "kube-api-access-mvpg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.570402 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.578139 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "301002c3-71f8-41bc-bf96-babd19cc994c" (UID: "301002c3-71f8-41bc-bf96-babd19cc994c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.586943 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "301002c3-71f8-41bc-bf96-babd19cc994c" (UID: "301002c3-71f8-41bc-bf96-babd19cc994c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.588144 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.593232 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-config" (OuterVolumeSpecName: "config") pod "301002c3-71f8-41bc-bf96-babd19cc994c" (UID: "301002c3-71f8-41bc-bf96-babd19cc994c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.605265 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "301002c3-71f8-41bc-bf96-babd19cc994c" (UID: "301002c3-71f8-41bc-bf96-babd19cc994c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.612644 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.626155 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-config-data" (OuterVolumeSpecName: "config-data") pod "286dd62a-ea0d-4d35-ac51-ff814de3eef0" (UID: "286dd62a-ea0d-4d35-ac51-ff814de3eef0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635220 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkdbb\" (UniqueName: \"kubernetes.io/projected/301002c3-71f8-41bc-bf96-babd19cc994c-kube-api-access-mkdbb\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635255 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635264 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635299 5035 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635309 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635318 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635327 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635334 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/286dd62a-ea0d-4d35-ac51-ff814de3eef0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635342 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvpg8\" (UniqueName: \"kubernetes.io/projected/286dd62a-ea0d-4d35-ac51-ff814de3eef0-kube-api-access-mvpg8\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635370 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635378 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/301002c3-71f8-41bc-bf96-babd19cc994c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635386 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:26 crc kubenswrapper[5035]: I1124 19:35:26.635394 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/286dd62a-ea0d-4d35-ac51-ff814de3eef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.390838 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7bbc55bc-tf5sn" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.391109 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.432154 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f7bbc55bc-tf5sn"] Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.440386 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f7bbc55bc-tf5sn"] Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.448978 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.460795 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.469810 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:27 crc kubenswrapper[5035]: E1124 19:35:27.470184 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="301002c3-71f8-41bc-bf96-babd19cc994c" containerName="dnsmasq-dns" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470204 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="301002c3-71f8-41bc-bf96-babd19cc994c" containerName="dnsmasq-dns" Nov 24 19:35:27 crc kubenswrapper[5035]: E1124 19:35:27.470217 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-central-agent" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470225 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-central-agent" Nov 24 19:35:27 crc kubenswrapper[5035]: E1124 19:35:27.470246 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-notification-agent" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470254 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-notification-agent" Nov 24 19:35:27 crc kubenswrapper[5035]: E1124 19:35:27.470272 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="301002c3-71f8-41bc-bf96-babd19cc994c" containerName="init" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470279 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="301002c3-71f8-41bc-bf96-babd19cc994c" containerName="init" Nov 24 19:35:27 crc kubenswrapper[5035]: E1124 19:35:27.470308 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="sg-core" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470315 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="sg-core" Nov 24 19:35:27 crc kubenswrapper[5035]: E1124 19:35:27.470337 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="proxy-httpd" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470345 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="proxy-httpd" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470500 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="proxy-httpd" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470518 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="sg-core" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470531 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="301002c3-71f8-41bc-bf96-babd19cc994c" containerName="dnsmasq-dns" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470542 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-central-agent" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.470550 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" containerName="ceilometer-notification-agent" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.472851 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.475242 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.475439 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.479032 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.490972 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.562770 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.562843 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.562950 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-config-data\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.563011 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-run-httpd\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.563032 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-log-httpd\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.563129 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.563177 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-scripts\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.563219 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgrbt\" (UniqueName: \"kubernetes.io/projected/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-kube-api-access-bgrbt\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.665274 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.665457 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-scripts\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.665556 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgrbt\" (UniqueName: \"kubernetes.io/projected/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-kube-api-access-bgrbt\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.665608 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.665672 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.665764 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-config-data\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.666763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-run-httpd\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.666909 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-log-httpd\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.667168 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-run-httpd\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.667198 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-log-httpd\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.669156 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.669251 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.674268 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-scripts\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.676665 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.686387 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-config-data\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.696821 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgrbt\" (UniqueName: \"kubernetes.io/projected/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-kube-api-access-bgrbt\") pod \"ceilometer-0\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " pod="openstack/ceilometer-0" Nov 24 19:35:27 crc kubenswrapper[5035]: I1124 19:35:27.789809 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 19:35:28 crc kubenswrapper[5035]: I1124 19:35:28.217860 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="286dd62a-ea0d-4d35-ac51-ff814de3eef0" path="/var/lib/kubelet/pods/286dd62a-ea0d-4d35-ac51-ff814de3eef0/volumes" Nov 24 19:35:28 crc kubenswrapper[5035]: I1124 19:35:28.219063 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="301002c3-71f8-41bc-bf96-babd19cc994c" path="/var/lib/kubelet/pods/301002c3-71f8-41bc-bf96-babd19cc994c/volumes" Nov 24 19:35:28 crc kubenswrapper[5035]: I1124 19:35:28.254496 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 19:35:28 crc kubenswrapper[5035]: W1124 19:35:28.254906 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d9647e3_bf7f_40db_86c8_299b2b08d5dd.slice/crio-a81d0ac1531422683a37f5eb4d01ea0beac26e0e3deb60170dbca0dacc9cd16a WatchSource:0}: Error finding container a81d0ac1531422683a37f5eb4d01ea0beac26e0e3deb60170dbca0dacc9cd16a: Status 404 returned error can't find the container with id a81d0ac1531422683a37f5eb4d01ea0beac26e0e3deb60170dbca0dacc9cd16a Nov 24 19:35:28 crc kubenswrapper[5035]: I1124 19:35:28.409352 5035 generic.go:334] "Generic (PLEG): container finished" podID="99e36d14-3aed-4554-8cab-f20fa21a29d8" containerID="b2a4c21951ea187c9e40b0b6d4b3ce1c29e72c640188b0c8eeed8acf139c21ce" exitCode=0 Nov 24 19:35:28 crc kubenswrapper[5035]: I1124 19:35:28.409412 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gnkh9" event={"ID":"99e36d14-3aed-4554-8cab-f20fa21a29d8","Type":"ContainerDied","Data":"b2a4c21951ea187c9e40b0b6d4b3ce1c29e72c640188b0c8eeed8acf139c21ce"} Nov 24 19:35:28 crc kubenswrapper[5035]: I1124 19:35:28.411532 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerStarted","Data":"a81d0ac1531422683a37f5eb4d01ea0beac26e0e3deb60170dbca0dacc9cd16a"} Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.421457 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerStarted","Data":"38101215349c39cf47ca984ecf78609bd7b92ee341d3a7a8bf98ddee7cb3d1bf"} Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.763537 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.815807 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-combined-ca-bundle\") pod \"99e36d14-3aed-4554-8cab-f20fa21a29d8\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.815908 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-config-data\") pod \"99e36d14-3aed-4554-8cab-f20fa21a29d8\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.815980 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79jrn\" (UniqueName: \"kubernetes.io/projected/99e36d14-3aed-4554-8cab-f20fa21a29d8-kube-api-access-79jrn\") pod \"99e36d14-3aed-4554-8cab-f20fa21a29d8\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.816003 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-scripts\") pod \"99e36d14-3aed-4554-8cab-f20fa21a29d8\" (UID: \"99e36d14-3aed-4554-8cab-f20fa21a29d8\") " Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.821352 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-scripts" (OuterVolumeSpecName: "scripts") pod "99e36d14-3aed-4554-8cab-f20fa21a29d8" (UID: "99e36d14-3aed-4554-8cab-f20fa21a29d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.821398 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e36d14-3aed-4554-8cab-f20fa21a29d8-kube-api-access-79jrn" (OuterVolumeSpecName: "kube-api-access-79jrn") pod "99e36d14-3aed-4554-8cab-f20fa21a29d8" (UID: "99e36d14-3aed-4554-8cab-f20fa21a29d8"). InnerVolumeSpecName "kube-api-access-79jrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.842217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-config-data" (OuterVolumeSpecName: "config-data") pod "99e36d14-3aed-4554-8cab-f20fa21a29d8" (UID: "99e36d14-3aed-4554-8cab-f20fa21a29d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.846022 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99e36d14-3aed-4554-8cab-f20fa21a29d8" (UID: "99e36d14-3aed-4554-8cab-f20fa21a29d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.921511 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79jrn\" (UniqueName: \"kubernetes.io/projected/99e36d14-3aed-4554-8cab-f20fa21a29d8-kube-api-access-79jrn\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.921783 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.921860 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:29 crc kubenswrapper[5035]: I1124 19:35:29.921940 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e36d14-3aed-4554-8cab-f20fa21a29d8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.432798 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gnkh9" event={"ID":"99e36d14-3aed-4554-8cab-f20fa21a29d8","Type":"ContainerDied","Data":"98a21a092692ecdd8e983256d4e18629ca6b622ad53cdecfd2b7efa418ad0b1e"} Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.436503 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98a21a092692ecdd8e983256d4e18629ca6b622ad53cdecfd2b7efa418ad0b1e" Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.436531 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerStarted","Data":"335c8dcab2716763f8973d61075470d1db17488e9d6e11a991abe8e82e7cf7b2"} Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.436549 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerStarted","Data":"16643cedcf8ef7af098ed9145fb085b98d72a5c98bf66401e0377161fcd54950"} Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.433078 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gnkh9" Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.695518 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.695930 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-log" containerID="cri-o://160a7bfaec1325b20eb24874f945f585a3af874dfd87562a5b8164bdd9d6d7e0" gracePeriod=30 Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.696581 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-api" containerID="cri-o://059cc1ecdc7b970cadee29e2db971d102d536c8312737643c621bc251aae3f34" gracePeriod=30 Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.774454 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.774728 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1a65c0b9-7094-4f6c-85af-6137ccf3d944" containerName="nova-scheduler-scheduler" containerID="cri-o://a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" gracePeriod=30 Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.794372 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.794617 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-log" containerID="cri-o://b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858" gracePeriod=30 Nov 24 19:35:30 crc kubenswrapper[5035]: I1124 19:35:30.795062 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-metadata" containerID="cri-o://759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327" gracePeriod=30 Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.455255 5035 generic.go:334] "Generic (PLEG): container finished" podID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerID="b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858" exitCode=143 Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.455673 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6fed84b-a0cc-4bae-84af-f318b03d032e","Type":"ContainerDied","Data":"b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858"} Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.458996 5035 generic.go:334] "Generic (PLEG): container finished" podID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerID="059cc1ecdc7b970cadee29e2db971d102d536c8312737643c621bc251aae3f34" exitCode=0 Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.459019 5035 generic.go:334] "Generic (PLEG): container finished" podID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerID="160a7bfaec1325b20eb24874f945f585a3af874dfd87562a5b8164bdd9d6d7e0" exitCode=143 Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.459036 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f","Type":"ContainerDied","Data":"059cc1ecdc7b970cadee29e2db971d102d536c8312737643c621bc251aae3f34"} Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.459053 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f","Type":"ContainerDied","Data":"160a7bfaec1325b20eb24874f945f585a3af874dfd87562a5b8164bdd9d6d7e0"} Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.459062 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f","Type":"ContainerDied","Data":"a239b0fda47def74fc0c7e6693b7d5ec0359812de2c1338baaaaa1ca19692cb9"} Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.459070 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a239b0fda47def74fc0c7e6693b7d5ec0359812de2c1338baaaaa1ca19692cb9" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.575048 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:31 crc kubenswrapper[5035]: E1124 19:35:31.676226 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 19:35:31 crc kubenswrapper[5035]: E1124 19:35:31.678343 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 19:35:31 crc kubenswrapper[5035]: E1124 19:35:31.680154 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 19:35:31 crc kubenswrapper[5035]: E1124 19:35:31.680308 5035 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1a65c0b9-7094-4f6c-85af-6137ccf3d944" containerName="nova-scheduler-scheduler" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.693092 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-config-data\") pod \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.693193 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-combined-ca-bundle\") pod \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.693250 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-logs\") pod \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.693334 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-public-tls-certs\") pod \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.693436 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-internal-tls-certs\") pod \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.693530 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nw6v\" (UniqueName: \"kubernetes.io/projected/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-kube-api-access-8nw6v\") pod \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\" (UID: \"f83427d4-5ef3-479b-8ed1-e1cffbea5c3f\") " Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.694210 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-logs" (OuterVolumeSpecName: "logs") pod "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" (UID: "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.709178 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-kube-api-access-8nw6v" (OuterVolumeSpecName: "kube-api-access-8nw6v") pod "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" (UID: "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f"). InnerVolumeSpecName "kube-api-access-8nw6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.725383 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" (UID: "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.725515 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-config-data" (OuterVolumeSpecName: "config-data") pod "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" (UID: "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.757227 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" (UID: "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.762285 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" (UID: "f83427d4-5ef3-479b-8ed1-e1cffbea5c3f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.795509 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.795575 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.795591 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.795603 5035 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.795615 5035 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:31 crc kubenswrapper[5035]: I1124 19:35:31.795627 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nw6v\" (UniqueName: \"kubernetes.io/projected/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f-kube-api-access-8nw6v\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.470285 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.471657 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerStarted","Data":"e4c3884585d3e48016841fb56e521726b56a30dba594469d70c110ca1cefe0c3"} Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.471816 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.496754 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.324787663 podStartE2EDuration="5.49673722s" podCreationTimestamp="2025-11-24 19:35:27 +0000 UTC" firstStartedPulling="2025-11-24 19:35:28.257546461 +0000 UTC m=+1266.780052718" lastFinishedPulling="2025-11-24 19:35:31.429496018 +0000 UTC m=+1269.952002275" observedRunningTime="2025-11-24 19:35:32.493716057 +0000 UTC m=+1271.016222314" watchObservedRunningTime="2025-11-24 19:35:32.49673722 +0000 UTC m=+1271.019243477" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.513005 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.532029 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.540569 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:32 crc kubenswrapper[5035]: E1124 19:35:32.540895 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e36d14-3aed-4554-8cab-f20fa21a29d8" containerName="nova-manage" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.540914 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e36d14-3aed-4554-8cab-f20fa21a29d8" containerName="nova-manage" Nov 24 19:35:32 crc kubenswrapper[5035]: E1124 19:35:32.540949 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-api" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.540955 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-api" Nov 24 19:35:32 crc kubenswrapper[5035]: E1124 19:35:32.540974 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-log" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.540979 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-log" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.541141 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-log" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.541158 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" containerName="nova-api-api" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.541167 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e36d14-3aed-4554-8cab-f20fa21a29d8" containerName="nova-manage" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.542049 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.544257 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.544576 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.544601 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.571840 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.610913 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks8sn\" (UniqueName: \"kubernetes.io/projected/2f870b4b-afbf-48bc-95d0-ba675cffaa49-kube-api-access-ks8sn\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.610998 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-config-data\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.611183 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f870b4b-afbf-48bc-95d0-ba675cffaa49-logs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.611231 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.611272 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.611370 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.712926 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-config-data\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.713003 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f870b4b-afbf-48bc-95d0-ba675cffaa49-logs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.713024 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.713048 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.713080 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.713127 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks8sn\" (UniqueName: \"kubernetes.io/projected/2f870b4b-afbf-48bc-95d0-ba675cffaa49-kube-api-access-ks8sn\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.714048 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f870b4b-afbf-48bc-95d0-ba675cffaa49-logs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.719621 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.731739 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.731796 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-public-tls-certs\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.732065 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f870b4b-afbf-48bc-95d0-ba675cffaa49-config-data\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.735240 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks8sn\" (UniqueName: \"kubernetes.io/projected/2f870b4b-afbf-48bc-95d0-ba675cffaa49-kube-api-access-ks8sn\") pod \"nova-api-0\" (UID: \"2f870b4b-afbf-48bc-95d0-ba675cffaa49\") " pod="openstack/nova-api-0" Nov 24 19:35:32 crc kubenswrapper[5035]: I1124 19:35:32.855133 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 19:35:33 crc kubenswrapper[5035]: I1124 19:35:33.312213 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 19:35:33 crc kubenswrapper[5035]: W1124 19:35:33.312279 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f870b4b_afbf_48bc_95d0_ba675cffaa49.slice/crio-b340fa7996788ef2fe7514638fe312b1b676a305ce6e55d2e993dc115e851643 WatchSource:0}: Error finding container b340fa7996788ef2fe7514638fe312b1b676a305ce6e55d2e993dc115e851643: Status 404 returned error can't find the container with id b340fa7996788ef2fe7514638fe312b1b676a305ce6e55d2e993dc115e851643 Nov 24 19:35:33 crc kubenswrapper[5035]: I1124 19:35:33.480431 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f870b4b-afbf-48bc-95d0-ba675cffaa49","Type":"ContainerStarted","Data":"b340fa7996788ef2fe7514638fe312b1b676a305ce6e55d2e993dc115e851643"} Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.022305 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": read tcp 10.217.0.2:36136->10.217.0.183:8775: read: connection reset by peer" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.022414 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": read tcp 10.217.0.2:36146->10.217.0.183:8775: read: connection reset by peer" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.214389 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f83427d4-5ef3-479b-8ed1-e1cffbea5c3f" path="/var/lib/kubelet/pods/f83427d4-5ef3-479b-8ed1-e1cffbea5c3f/volumes" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.369364 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.448905 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8swr\" (UniqueName: \"kubernetes.io/projected/f6fed84b-a0cc-4bae-84af-f318b03d032e-kube-api-access-r8swr\") pod \"f6fed84b-a0cc-4bae-84af-f318b03d032e\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.448963 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-nova-metadata-tls-certs\") pod \"f6fed84b-a0cc-4bae-84af-f318b03d032e\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.449015 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6fed84b-a0cc-4bae-84af-f318b03d032e-logs\") pod \"f6fed84b-a0cc-4bae-84af-f318b03d032e\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.449039 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-combined-ca-bundle\") pod \"f6fed84b-a0cc-4bae-84af-f318b03d032e\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.449129 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-config-data\") pod \"f6fed84b-a0cc-4bae-84af-f318b03d032e\" (UID: \"f6fed84b-a0cc-4bae-84af-f318b03d032e\") " Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.449991 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6fed84b-a0cc-4bae-84af-f318b03d032e-logs" (OuterVolumeSpecName: "logs") pod "f6fed84b-a0cc-4bae-84af-f318b03d032e" (UID: "f6fed84b-a0cc-4bae-84af-f318b03d032e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.465343 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6fed84b-a0cc-4bae-84af-f318b03d032e-kube-api-access-r8swr" (OuterVolumeSpecName: "kube-api-access-r8swr") pod "f6fed84b-a0cc-4bae-84af-f318b03d032e" (UID: "f6fed84b-a0cc-4bae-84af-f318b03d032e"). InnerVolumeSpecName "kube-api-access-r8swr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.492771 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-config-data" (OuterVolumeSpecName: "config-data") pod "f6fed84b-a0cc-4bae-84af-f318b03d032e" (UID: "f6fed84b-a0cc-4bae-84af-f318b03d032e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.494106 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6fed84b-a0cc-4bae-84af-f318b03d032e" (UID: "f6fed84b-a0cc-4bae-84af-f318b03d032e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.510792 5035 generic.go:334] "Generic (PLEG): container finished" podID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerID="759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327" exitCode=0 Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.510870 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6fed84b-a0cc-4bae-84af-f318b03d032e","Type":"ContainerDied","Data":"759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327"} Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.510903 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6fed84b-a0cc-4bae-84af-f318b03d032e","Type":"ContainerDied","Data":"058ddd22913d1254051827fa9d6388c150e56395d388b09c51e31df10fc0af6e"} Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.510924 5035 scope.go:117] "RemoveContainer" containerID="759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.511072 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.514875 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f870b4b-afbf-48bc-95d0-ba675cffaa49","Type":"ContainerStarted","Data":"4b7abfab746fa0220ae83ae8c4444134304dcd47a99834e378025195114599fe"} Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.515205 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f870b4b-afbf-48bc-95d0-ba675cffaa49","Type":"ContainerStarted","Data":"1862e5c08ccc5682cc316df14d8eb599f4a467bc3e36543b81f1d3ad2fa96b9b"} Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.538433 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f6fed84b-a0cc-4bae-84af-f318b03d032e" (UID: "f6fed84b-a0cc-4bae-84af-f318b03d032e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.552572 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8swr\" (UniqueName: \"kubernetes.io/projected/f6fed84b-a0cc-4bae-84af-f318b03d032e-kube-api-access-r8swr\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.552598 5035 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.552609 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6fed84b-a0cc-4bae-84af-f318b03d032e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.552618 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.552626 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6fed84b-a0cc-4bae-84af-f318b03d032e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.584307 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.584262819 podStartE2EDuration="2.584262819s" podCreationTimestamp="2025-11-24 19:35:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:35:34.566001463 +0000 UTC m=+1273.088507720" watchObservedRunningTime="2025-11-24 19:35:34.584262819 +0000 UTC m=+1273.106769076" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.608593 5035 scope.go:117] "RemoveContainer" containerID="b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.639817 5035 scope.go:117] "RemoveContainer" containerID="759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327" Nov 24 19:35:34 crc kubenswrapper[5035]: E1124 19:35:34.640487 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327\": container with ID starting with 759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327 not found: ID does not exist" containerID="759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.640543 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327"} err="failed to get container status \"759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327\": rpc error: code = NotFound desc = could not find container \"759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327\": container with ID starting with 759a81d33c80f8ec81cfe42287645972cc6dcff0f180234aa88d9b1685b0f327 not found: ID does not exist" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.640577 5035 scope.go:117] "RemoveContainer" containerID="b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858" Nov 24 19:35:34 crc kubenswrapper[5035]: E1124 19:35:34.640895 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858\": container with ID starting with b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858 not found: ID does not exist" containerID="b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.640927 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858"} err="failed to get container status \"b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858\": rpc error: code = NotFound desc = could not find container \"b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858\": container with ID starting with b5a8beb9e93a7b0aedf90be0d7298c2a25d22f76cdfd0ee8e98ca5f6e6452858 not found: ID does not exist" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.839979 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.849670 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.865617 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:35:34 crc kubenswrapper[5035]: E1124 19:35:34.866077 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-metadata" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.866100 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-metadata" Nov 24 19:35:34 crc kubenswrapper[5035]: E1124 19:35:34.866137 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-log" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.866146 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-log" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.866394 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-log" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.866419 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" containerName="nova-metadata-metadata" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.870499 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.877371 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.877909 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.894204 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.958624 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95z2r\" (UniqueName: \"kubernetes.io/projected/5e338747-46ee-4e25-a2cc-6696a162487a-kube-api-access-95z2r\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.958687 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.958796 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.958851 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e338747-46ee-4e25-a2cc-6696a162487a-logs\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:34 crc kubenswrapper[5035]: I1124 19:35:34.958888 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-config-data\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.060978 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.061043 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e338747-46ee-4e25-a2cc-6696a162487a-logs\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.061077 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-config-data\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.061230 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95z2r\" (UniqueName: \"kubernetes.io/projected/5e338747-46ee-4e25-a2cc-6696a162487a-kube-api-access-95z2r\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.061263 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.061483 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e338747-46ee-4e25-a2cc-6696a162487a-logs\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.065748 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.066438 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-config-data\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.073987 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e338747-46ee-4e25-a2cc-6696a162487a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.080348 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95z2r\" (UniqueName: \"kubernetes.io/projected/5e338747-46ee-4e25-a2cc-6696a162487a-kube-api-access-95z2r\") pod \"nova-metadata-0\" (UID: \"5e338747-46ee-4e25-a2cc-6696a162487a\") " pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.194695 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 19:35:35 crc kubenswrapper[5035]: I1124 19:35:35.970597 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.163734 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.215502 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6fed84b-a0cc-4bae-84af-f318b03d032e" path="/var/lib/kubelet/pods/f6fed84b-a0cc-4bae-84af-f318b03d032e/volumes" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.283821 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-config-data\") pod \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.284239 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwj8z\" (UniqueName: \"kubernetes.io/projected/1a65c0b9-7094-4f6c-85af-6137ccf3d944-kube-api-access-jwj8z\") pod \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.284366 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-combined-ca-bundle\") pod \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\" (UID: \"1a65c0b9-7094-4f6c-85af-6137ccf3d944\") " Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.290572 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a65c0b9-7094-4f6c-85af-6137ccf3d944-kube-api-access-jwj8z" (OuterVolumeSpecName: "kube-api-access-jwj8z") pod "1a65c0b9-7094-4f6c-85af-6137ccf3d944" (UID: "1a65c0b9-7094-4f6c-85af-6137ccf3d944"). InnerVolumeSpecName "kube-api-access-jwj8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.319186 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-config-data" (OuterVolumeSpecName: "config-data") pod "1a65c0b9-7094-4f6c-85af-6137ccf3d944" (UID: "1a65c0b9-7094-4f6c-85af-6137ccf3d944"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.348589 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a65c0b9-7094-4f6c-85af-6137ccf3d944" (UID: "1a65c0b9-7094-4f6c-85af-6137ccf3d944"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.386475 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.386516 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a65c0b9-7094-4f6c-85af-6137ccf3d944-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.386528 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwj8z\" (UniqueName: \"kubernetes.io/projected/1a65c0b9-7094-4f6c-85af-6137ccf3d944-kube-api-access-jwj8z\") on node \"crc\" DevicePath \"\"" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.533497 5035 generic.go:334] "Generic (PLEG): container finished" podID="1a65c0b9-7094-4f6c-85af-6137ccf3d944" containerID="a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" exitCode=0 Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.533557 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a65c0b9-7094-4f6c-85af-6137ccf3d944","Type":"ContainerDied","Data":"a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e"} Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.533568 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.533584 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a65c0b9-7094-4f6c-85af-6137ccf3d944","Type":"ContainerDied","Data":"0eea632520bb9e31969bb0b3a2637a80895f25541638cefc19720dfc3396197c"} Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.533621 5035 scope.go:117] "RemoveContainer" containerID="a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.536535 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e338747-46ee-4e25-a2cc-6696a162487a","Type":"ContainerStarted","Data":"bee0abc7cf402a184d291ec30a0feaf311e21c083036b597c9232360e8325a05"} Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.536574 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e338747-46ee-4e25-a2cc-6696a162487a","Type":"ContainerStarted","Data":"7562a37197e80f55558d415038c215a90158aca2233093515fecbd397686c350"} Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.536593 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e338747-46ee-4e25-a2cc-6696a162487a","Type":"ContainerStarted","Data":"8a3d372fa924d42bdd5c248f4705cbba0e76193bd7354f493aa71a9dda4c0c56"} Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.561494 5035 scope.go:117] "RemoveContainer" containerID="a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" Nov 24 19:35:36 crc kubenswrapper[5035]: E1124 19:35:36.561907 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e\": container with ID starting with a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e not found: ID does not exist" containerID="a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.561938 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e"} err="failed to get container status \"a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e\": rpc error: code = NotFound desc = could not find container \"a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e\": container with ID starting with a851fe70936b3e187a1af9818634d493f6c80f585e6655c0fa2cd6b4bf6c319e not found: ID does not exist" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.605384 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.622608 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.629484 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:35:36 crc kubenswrapper[5035]: E1124 19:35:36.629894 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a65c0b9-7094-4f6c-85af-6137ccf3d944" containerName="nova-scheduler-scheduler" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.629912 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a65c0b9-7094-4f6c-85af-6137ccf3d944" containerName="nova-scheduler-scheduler" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.630092 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a65c0b9-7094-4f6c-85af-6137ccf3d944" containerName="nova-scheduler-scheduler" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.633135 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.640367 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.647985 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.692349 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd81445-025b-4359-8783-2da4b4181807-config-data\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.692441 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd81445-025b-4359-8783-2da4b4181807-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.692710 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2rch\" (UniqueName: \"kubernetes.io/projected/5bd81445-025b-4359-8783-2da4b4181807-kube-api-access-t2rch\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.795121 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2rch\" (UniqueName: \"kubernetes.io/projected/5bd81445-025b-4359-8783-2da4b4181807-kube-api-access-t2rch\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.795198 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd81445-025b-4359-8783-2da4b4181807-config-data\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.795269 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd81445-025b-4359-8783-2da4b4181807-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.799605 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bd81445-025b-4359-8783-2da4b4181807-config-data\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.811096 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bd81445-025b-4359-8783-2da4b4181807-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.811131 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2rch\" (UniqueName: \"kubernetes.io/projected/5bd81445-025b-4359-8783-2da4b4181807-kube-api-access-t2rch\") pod \"nova-scheduler-0\" (UID: \"5bd81445-025b-4359-8783-2da4b4181807\") " pod="openstack/nova-scheduler-0" Nov 24 19:35:36 crc kubenswrapper[5035]: I1124 19:35:36.964790 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 19:35:37 crc kubenswrapper[5035]: W1124 19:35:37.426690 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bd81445_025b_4359_8783_2da4b4181807.slice/crio-cb85aa4d745de998e9d2a40e009c3e74f822d41e5d8da0fb313c7948a4851098 WatchSource:0}: Error finding container cb85aa4d745de998e9d2a40e009c3e74f822d41e5d8da0fb313c7948a4851098: Status 404 returned error can't find the container with id cb85aa4d745de998e9d2a40e009c3e74f822d41e5d8da0fb313c7948a4851098 Nov 24 19:35:37 crc kubenswrapper[5035]: I1124 19:35:37.430737 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 19:35:37 crc kubenswrapper[5035]: I1124 19:35:37.546195 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5bd81445-025b-4359-8783-2da4b4181807","Type":"ContainerStarted","Data":"cb85aa4d745de998e9d2a40e009c3e74f822d41e5d8da0fb313c7948a4851098"} Nov 24 19:35:37 crc kubenswrapper[5035]: I1124 19:35:37.568465 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.568439093 podStartE2EDuration="3.568439093s" podCreationTimestamp="2025-11-24 19:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:35:37.562129491 +0000 UTC m=+1276.084635748" watchObservedRunningTime="2025-11-24 19:35:37.568439093 +0000 UTC m=+1276.090945370" Nov 24 19:35:38 crc kubenswrapper[5035]: I1124 19:35:38.209498 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a65c0b9-7094-4f6c-85af-6137ccf3d944" path="/var/lib/kubelet/pods/1a65c0b9-7094-4f6c-85af-6137ccf3d944/volumes" Nov 24 19:35:38 crc kubenswrapper[5035]: I1124 19:35:38.557962 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5bd81445-025b-4359-8783-2da4b4181807","Type":"ContainerStarted","Data":"fd4994d7234331e3e463786d28c21466eb1de36aab1b21e81f4a72d43e3b814b"} Nov 24 19:35:38 crc kubenswrapper[5035]: I1124 19:35:38.593558 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5935341210000002 podStartE2EDuration="2.593534121s" podCreationTimestamp="2025-11-24 19:35:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:35:38.577068364 +0000 UTC m=+1277.099574631" watchObservedRunningTime="2025-11-24 19:35:38.593534121 +0000 UTC m=+1277.116040398" Nov 24 19:35:40 crc kubenswrapper[5035]: I1124 19:35:40.195692 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:35:40 crc kubenswrapper[5035]: I1124 19:35:40.196392 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 19:35:41 crc kubenswrapper[5035]: I1124 19:35:41.965167 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 19:35:42 crc kubenswrapper[5035]: I1124 19:35:42.855398 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 19:35:42 crc kubenswrapper[5035]: I1124 19:35:42.855725 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 19:35:43 crc kubenswrapper[5035]: I1124 19:35:43.867619 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f870b4b-afbf-48bc-95d0-ba675cffaa49" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 19:35:43 crc kubenswrapper[5035]: I1124 19:35:43.867648 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f870b4b-afbf-48bc-95d0-ba675cffaa49" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.192:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 19:35:45 crc kubenswrapper[5035]: I1124 19:35:45.196273 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 19:35:45 crc kubenswrapper[5035]: I1124 19:35:45.196640 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 19:35:46 crc kubenswrapper[5035]: I1124 19:35:46.208559 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e338747-46ee-4e25-a2cc-6696a162487a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 19:35:46 crc kubenswrapper[5035]: I1124 19:35:46.208556 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e338747-46ee-4e25-a2cc-6696a162487a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 19:35:46 crc kubenswrapper[5035]: I1124 19:35:46.966018 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 19:35:47 crc kubenswrapper[5035]: I1124 19:35:47.012329 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 19:35:47 crc kubenswrapper[5035]: I1124 19:35:47.680069 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 19:35:52 crc kubenswrapper[5035]: I1124 19:35:52.863662 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 19:35:52 crc kubenswrapper[5035]: I1124 19:35:52.864231 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 19:35:52 crc kubenswrapper[5035]: I1124 19:35:52.864592 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 19:35:52 crc kubenswrapper[5035]: I1124 19:35:52.864636 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 19:35:52 crc kubenswrapper[5035]: I1124 19:35:52.871846 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 19:35:52 crc kubenswrapper[5035]: I1124 19:35:52.878087 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 19:35:55 crc kubenswrapper[5035]: I1124 19:35:55.204614 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 19:35:55 crc kubenswrapper[5035]: I1124 19:35:55.211321 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 19:35:55 crc kubenswrapper[5035]: I1124 19:35:55.215305 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 19:35:55 crc kubenswrapper[5035]: I1124 19:35:55.727819 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 19:35:57 crc kubenswrapper[5035]: I1124 19:35:57.802078 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 19:36:06 crc kubenswrapper[5035]: I1124 19:36:06.901729 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:36:07 crc kubenswrapper[5035]: I1124 19:36:07.711330 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:36:10 crc kubenswrapper[5035]: I1124 19:36:10.967430 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerName="rabbitmq" containerID="cri-o://614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f" gracePeriod=604796 Nov 24 19:36:11 crc kubenswrapper[5035]: I1124 19:36:11.659399 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="8716e4c9-35b0-4922-afc2-13006d78283a" containerName="rabbitmq" containerID="cri-o://4977ad3c8b82328502e1b7cffa7822db6d743aec4c461b5644175f59a455814d" gracePeriod=604797 Nov 24 19:36:15 crc kubenswrapper[5035]: I1124 19:36:15.234287 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:36:15 crc kubenswrapper[5035]: I1124 19:36:15.234661 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.560759 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691003 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-tls\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691065 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691147 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-plugins-conf\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691177 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpqlq\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-kube-api-access-mpqlq\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691266 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-config-data\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691315 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-erlang-cookie\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-plugins\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691377 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-confd\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691445 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ed0e7cb-164f-4689-8714-1f11bfa25725-erlang-cookie-secret\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691469 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-server-conf\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.691506 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ed0e7cb-164f-4689-8714-1f11bfa25725-pod-info\") pod \"4ed0e7cb-164f-4689-8714-1f11bfa25725\" (UID: \"4ed0e7cb-164f-4689-8714-1f11bfa25725\") " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.696783 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.702316 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.702563 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.707975 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4ed0e7cb-164f-4689-8714-1f11bfa25725-pod-info" (OuterVolumeSpecName: "pod-info") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.710572 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ed0e7cb-164f-4689-8714-1f11bfa25725-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.715959 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-kube-api-access-mpqlq" (OuterVolumeSpecName: "kube-api-access-mpqlq") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "kube-api-access-mpqlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.730509 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.733454 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.770175 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-config-data" (OuterVolumeSpecName: "config-data") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796750 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796782 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796793 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796803 5035 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ed0e7cb-164f-4689-8714-1f11bfa25725-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796812 5035 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ed0e7cb-164f-4689-8714-1f11bfa25725-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796822 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796848 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796859 5035 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.796870 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpqlq\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-kube-api-access-mpqlq\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.865749 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.869606 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-server-conf" (OuterVolumeSpecName: "server-conf") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.898270 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.898419 5035 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ed0e7cb-164f-4689-8714-1f11bfa25725-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.936091 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4ed0e7cb-164f-4689-8714-1f11bfa25725" (UID: "4ed0e7cb-164f-4689-8714-1f11bfa25725"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.937763 5035 generic.go:334] "Generic (PLEG): container finished" podID="8716e4c9-35b0-4922-afc2-13006d78283a" containerID="4977ad3c8b82328502e1b7cffa7822db6d743aec4c461b5644175f59a455814d" exitCode=0 Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.937819 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8716e4c9-35b0-4922-afc2-13006d78283a","Type":"ContainerDied","Data":"4977ad3c8b82328502e1b7cffa7822db6d743aec4c461b5644175f59a455814d"} Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.939494 5035 generic.go:334] "Generic (PLEG): container finished" podID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerID="614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f" exitCode=0 Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.939523 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ed0e7cb-164f-4689-8714-1f11bfa25725","Type":"ContainerDied","Data":"614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f"} Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.939547 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ed0e7cb-164f-4689-8714-1f11bfa25725","Type":"ContainerDied","Data":"3d68d7f492531271729710ef1645bc1020b27d15ad7057c2b794bf8b00746c86"} Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.939561 5035 scope.go:117] "RemoveContainer" containerID="614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.939510 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.977861 5035 scope.go:117] "RemoveContainer" containerID="13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04" Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.986370 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.995906 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:36:17 crc kubenswrapper[5035]: I1124 19:36:17.999944 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ed0e7cb-164f-4689-8714-1f11bfa25725-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.004199 5035 scope.go:117] "RemoveContainer" containerID="614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f" Nov 24 19:36:18 crc kubenswrapper[5035]: E1124 19:36:18.004619 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f\": container with ID starting with 614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f not found: ID does not exist" containerID="614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.004763 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f"} err="failed to get container status \"614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f\": rpc error: code = NotFound desc = could not find container \"614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f\": container with ID starting with 614a834cb0d99513e78be02f5f26e3c8ecfad516aefff0a421fa212142f3640f not found: ID does not exist" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.004864 5035 scope.go:117] "RemoveContainer" containerID="13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04" Nov 24 19:36:18 crc kubenswrapper[5035]: E1124 19:36:18.005606 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04\": container with ID starting with 13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04 not found: ID does not exist" containerID="13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.005722 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04"} err="failed to get container status \"13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04\": rpc error: code = NotFound desc = could not find container \"13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04\": container with ID starting with 13d4fa214e456485b52160f2ea06f02ed5c807fcd7e7eff4fd30bbfc2071da04 not found: ID does not exist" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.006743 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:36:18 crc kubenswrapper[5035]: E1124 19:36:18.007204 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerName="rabbitmq" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.007227 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerName="rabbitmq" Nov 24 19:36:18 crc kubenswrapper[5035]: E1124 19:36:18.007267 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerName="setup-container" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.007276 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerName="setup-container" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.007513 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed0e7cb-164f-4689-8714-1f11bfa25725" containerName="rabbitmq" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.008704 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.016316 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.016576 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.016826 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-c9p2l" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.017019 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.017095 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.017275 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.017585 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.021491 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.202951 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qnbn\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-kube-api-access-2qnbn\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.203601 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.203638 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.203685 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-config-data\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.203775 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.203833 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.204045 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00618566-3b37-4b50-92d5-5f7dabf6a55b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.204086 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.204113 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.204218 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00618566-3b37-4b50-92d5-5f7dabf6a55b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.204352 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.229849 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ed0e7cb-164f-4689-8714-1f11bfa25725" path="/var/lib/kubelet/pods/4ed0e7cb-164f-4689-8714-1f11bfa25725/volumes" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.305894 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.306282 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qnbn\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-kube-api-access-2qnbn\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.306356 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.307127 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.307789 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.307832 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.307979 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-config-data\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.308033 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.308062 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.308142 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00618566-3b37-4b50-92d5-5f7dabf6a55b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.308174 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.308190 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.308248 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00618566-3b37-4b50-92d5-5f7dabf6a55b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.308836 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-config-data\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.309618 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.309859 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.322018 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.322266 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.322687 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00618566-3b37-4b50-92d5-5f7dabf6a55b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.322911 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00618566-3b37-4b50-92d5-5f7dabf6a55b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.326422 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qnbn\" (UniqueName: \"kubernetes.io/projected/00618566-3b37-4b50-92d5-5f7dabf6a55b-kube-api-access-2qnbn\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.326607 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00618566-3b37-4b50-92d5-5f7dabf6a55b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.357005 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"00618566-3b37-4b50-92d5-5f7dabf6a55b\") " pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.408203 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.409392 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-erlang-cookie\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.409510 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.409603 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-confd\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.409696 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rb26\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-kube-api-access-7rb26\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.409803 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-tls\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.409917 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-config-data\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.410037 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-plugins-conf\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.410142 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-plugins\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.410937 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.411552 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.413781 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.414969 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.415898 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.416937 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-kube-api-access-7rb26" (OuterVolumeSpecName: "kube-api-access-7rb26") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "kube-api-access-7rb26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.451317 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-config-data" (OuterVolumeSpecName: "config-data") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512046 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8716e4c9-35b0-4922-afc2-13006d78283a-erlang-cookie-secret\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512126 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-server-conf\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512181 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8716e4c9-35b0-4922-afc2-13006d78283a-pod-info\") pod \"8716e4c9-35b0-4922-afc2-13006d78283a\" (UID: \"8716e4c9-35b0-4922-afc2-13006d78283a\") " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512851 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512887 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512903 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rb26\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-kube-api-access-7rb26\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512914 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512928 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512939 5035 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.512951 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.516856 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8716e4c9-35b0-4922-afc2-13006d78283a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.518111 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8716e4c9-35b0-4922-afc2-13006d78283a-pod-info" (OuterVolumeSpecName: "pod-info") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.552711 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.565873 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.583255 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-server-conf" (OuterVolumeSpecName: "server-conf") pod "8716e4c9-35b0-4922-afc2-13006d78283a" (UID: "8716e4c9-35b0-4922-afc2-13006d78283a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.614345 5035 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8716e4c9-35b0-4922-afc2-13006d78283a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.614381 5035 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8716e4c9-35b0-4922-afc2-13006d78283a-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.614394 5035 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8716e4c9-35b0-4922-afc2-13006d78283a-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.614405 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.614417 5035 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8716e4c9-35b0-4922-afc2-13006d78283a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.635881 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.954259 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8716e4c9-35b0-4922-afc2-13006d78283a","Type":"ContainerDied","Data":"45d8dbb82f0c81105f2938ebaf8dd1ff4ac1a3a394bc0890bc245205a72fb6a6"} Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.954344 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.954582 5035 scope.go:117] "RemoveContainer" containerID="4977ad3c8b82328502e1b7cffa7822db6d743aec4c461b5644175f59a455814d" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.977933 5035 scope.go:117] "RemoveContainer" containerID="821d307dfb370f447c67d71297fcefc4e81c98065f7a613d6310b10ef3ad01a6" Nov 24 19:36:18 crc kubenswrapper[5035]: I1124 19:36:18.999460 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.019384 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.030665 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:36:19 crc kubenswrapper[5035]: E1124 19:36:19.031119 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8716e4c9-35b0-4922-afc2-13006d78283a" containerName="setup-container" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.031135 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8716e4c9-35b0-4922-afc2-13006d78283a" containerName="setup-container" Nov 24 19:36:19 crc kubenswrapper[5035]: E1124 19:36:19.031145 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8716e4c9-35b0-4922-afc2-13006d78283a" containerName="rabbitmq" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.031151 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8716e4c9-35b0-4922-afc2-13006d78283a" containerName="rabbitmq" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.031356 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8716e4c9-35b0-4922-afc2-13006d78283a" containerName="rabbitmq" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.032308 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.037029 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.037378 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.037647 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.037841 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.038052 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.038438 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-t7zgv" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.038558 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.059685 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.080487 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226083 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226138 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226157 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226176 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmtjx\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-kube-api-access-dmtjx\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226246 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226284 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226332 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226366 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226398 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226544 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.226604 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.328234 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.328707 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.328743 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.328778 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329164 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329311 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329346 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329383 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmtjx\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-kube-api-access-dmtjx\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329494 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329544 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329598 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.329603 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.330080 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.330221 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.330271 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.331068 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.332876 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.333211 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.334360 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.336870 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.349922 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmtjx\" (UniqueName: \"kubernetes.io/projected/6d6f3cf9-b4c1-48e1-880b-8812b7f4421c-kube-api-access-dmtjx\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.359849 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.368480 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.814082 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.965250 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00618566-3b37-4b50-92d5-5f7dabf6a55b","Type":"ContainerStarted","Data":"a27f3749f149805709b20e330a92c8ad8b2412294d1c61472f0acf25e173f990"} Nov 24 19:36:19 crc kubenswrapper[5035]: I1124 19:36:19.968237 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c","Type":"ContainerStarted","Data":"dc8a84d2bc3503da9aef997fb3965bdf789dc478fc2d7ef600424c33060eeb52"} Nov 24 19:36:20 crc kubenswrapper[5035]: I1124 19:36:20.215098 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8716e4c9-35b0-4922-afc2-13006d78283a" path="/var/lib/kubelet/pods/8716e4c9-35b0-4922-afc2-13006d78283a/volumes" Nov 24 19:36:20 crc kubenswrapper[5035]: I1124 19:36:20.977994 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00618566-3b37-4b50-92d5-5f7dabf6a55b","Type":"ContainerStarted","Data":"11c83ac000c4cfea2e982277a67e6a9e0bf6fda4204328d4fb6d8bdf4ac2357c"} Nov 24 19:36:21 crc kubenswrapper[5035]: I1124 19:36:21.991118 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c","Type":"ContainerStarted","Data":"c91cab0b50981d6d3cf9c6d236923303448c342ae9118b3a328fc3fb6b80ae19"} Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.669521 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64b6dd64c5-mrbkx"] Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.671955 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.675054 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.680124 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64b6dd64c5-mrbkx"] Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.786910 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc9n5\" (UniqueName: \"kubernetes.io/projected/5426a0ef-fa48-4a37-a7b7-98601bb592a1-kube-api-access-zc9n5\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.786991 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.787025 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-config\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.787081 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-dns-svc\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.787351 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-nb\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.787397 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-sb\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.888994 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-nb\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.889061 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-sb\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.889444 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc9n5\" (UniqueName: \"kubernetes.io/projected/5426a0ef-fa48-4a37-a7b7-98601bb592a1-kube-api-access-zc9n5\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.889626 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.889696 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-config\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.889805 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-dns-svc\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.890096 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-nb\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.890579 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.890660 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-config\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.890730 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-dns-svc\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.894896 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-sb\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.906881 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc9n5\" (UniqueName: \"kubernetes.io/projected/5426a0ef-fa48-4a37-a7b7-98601bb592a1-kube-api-access-zc9n5\") pod \"dnsmasq-dns-64b6dd64c5-mrbkx\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:27 crc kubenswrapper[5035]: I1124 19:36:27.995525 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.463662 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64b6dd64c5-mrbkx"] Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.769356 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87"] Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.770873 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.773631 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.773871 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.773942 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.774231 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.784186 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87"] Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.908284 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.908375 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.908413 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:28 crc kubenswrapper[5035]: I1124 19:36:28.908552 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p84kg\" (UniqueName: \"kubernetes.io/projected/82a1649c-9b91-4ed5-bff3-acddd2292b4f-kube-api-access-p84kg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.010332 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.010593 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.010718 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.010936 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p84kg\" (UniqueName: \"kubernetes.io/projected/82a1649c-9b91-4ed5-bff3-acddd2292b4f-kube-api-access-p84kg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.015055 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.015386 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.016650 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.029731 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p84kg\" (UniqueName: \"kubernetes.io/projected/82a1649c-9b91-4ed5-bff3-acddd2292b4f-kube-api-access-p84kg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-28n87\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.054115 5035 generic.go:334] "Generic (PLEG): container finished" podID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerID="c2e1cd285bfe278236bc95b13d43a8aeaa7477123afdd6e34d0e98da412beb22" exitCode=0 Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.054169 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" event={"ID":"5426a0ef-fa48-4a37-a7b7-98601bb592a1","Type":"ContainerDied","Data":"c2e1cd285bfe278236bc95b13d43a8aeaa7477123afdd6e34d0e98da412beb22"} Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.054200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" event={"ID":"5426a0ef-fa48-4a37-a7b7-98601bb592a1","Type":"ContainerStarted","Data":"675b72838396d7fec753d17194ad4b64aaa8be7060317119391e6f1e7cd3f739"} Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.117452 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.672327 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87"] Nov 24 19:36:29 crc kubenswrapper[5035]: W1124 19:36:29.676692 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82a1649c_9b91_4ed5_bff3_acddd2292b4f.slice/crio-fb50b2681360f22c6f94d314b30405fdf2e853223676a4d6a6f011e8ecb89dcb WatchSource:0}: Error finding container fb50b2681360f22c6f94d314b30405fdf2e853223676a4d6a6f011e8ecb89dcb: Status 404 returned error can't find the container with id fb50b2681360f22c6f94d314b30405fdf2e853223676a4d6a6f011e8ecb89dcb Nov 24 19:36:29 crc kubenswrapper[5035]: I1124 19:36:29.686617 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 19:36:30 crc kubenswrapper[5035]: I1124 19:36:30.064337 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" event={"ID":"5426a0ef-fa48-4a37-a7b7-98601bb592a1","Type":"ContainerStarted","Data":"5b2b0e6e3f5ceee94ac8b82d7f324af5305cce4649b182cf54ebef3ede9f6b02"} Nov 24 19:36:30 crc kubenswrapper[5035]: I1124 19:36:30.064523 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:30 crc kubenswrapper[5035]: I1124 19:36:30.065774 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" event={"ID":"82a1649c-9b91-4ed5-bff3-acddd2292b4f","Type":"ContainerStarted","Data":"fb50b2681360f22c6f94d314b30405fdf2e853223676a4d6a6f011e8ecb89dcb"} Nov 24 19:36:30 crc kubenswrapper[5035]: I1124 19:36:30.089729 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" podStartSLOduration=3.089710253 podStartE2EDuration="3.089710253s" podCreationTimestamp="2025-11-24 19:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:36:30.082720373 +0000 UTC m=+1328.605226650" watchObservedRunningTime="2025-11-24 19:36:30.089710253 +0000 UTC m=+1328.612216520" Nov 24 19:36:37 crc kubenswrapper[5035]: I1124 19:36:37.997559 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.060014 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f95c456cf-dz6jb"] Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.060531 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" podUID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerName="dnsmasq-dns" containerID="cri-o://93b09cd18ef0bac4192a0c4d0ae1a671fd2cfc5fd437e2653f4e0176ad86ee34" gracePeriod=10 Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.216071 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c58867b6c-btg8h"] Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.217922 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c58867b6c-btg8h"] Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.218029 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.325915 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67bbz\" (UniqueName: \"kubernetes.io/projected/268a0d3d-f4b4-49c2-ae77-9ec588442031-kube-api-access-67bbz\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.326126 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-config\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.326173 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-openstack-edpm-ipam\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.326190 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-dns-svc\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.326226 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-sb\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.326254 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-nb\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.427783 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67bbz\" (UniqueName: \"kubernetes.io/projected/268a0d3d-f4b4-49c2-ae77-9ec588442031-kube-api-access-67bbz\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.427871 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-config\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.427901 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-openstack-edpm-ipam\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.427918 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-dns-svc\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.427939 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-sb\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.427958 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-nb\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.428928 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-nb\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.428969 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-config\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.429537 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-sb\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.429764 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-openstack-edpm-ipam\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.430975 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-dns-svc\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.447369 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67bbz\" (UniqueName: \"kubernetes.io/projected/268a0d3d-f4b4-49c2-ae77-9ec588442031-kube-api-access-67bbz\") pod \"dnsmasq-dns-c58867b6c-btg8h\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:38 crc kubenswrapper[5035]: I1124 19:36:38.589824 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.154186 5035 generic.go:334] "Generic (PLEG): container finished" podID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerID="93b09cd18ef0bac4192a0c4d0ae1a671fd2cfc5fd437e2653f4e0176ad86ee34" exitCode=0 Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.154275 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" event={"ID":"0e10dfb5-71c9-4b4c-a520-0845163d3b42","Type":"ContainerDied","Data":"93b09cd18ef0bac4192a0c4d0ae1a671fd2cfc5fd437e2653f4e0176ad86ee34"} Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.414145 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.551853 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-config\") pod \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.552228 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-dns-svc\") pod \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.552256 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-sb\") pod \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.552310 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg7wd\" (UniqueName: \"kubernetes.io/projected/0e10dfb5-71c9-4b4c-a520-0845163d3b42-kube-api-access-cg7wd\") pod \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.552541 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-nb\") pod \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\" (UID: \"0e10dfb5-71c9-4b4c-a520-0845163d3b42\") " Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.562485 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e10dfb5-71c9-4b4c-a520-0845163d3b42-kube-api-access-cg7wd" (OuterVolumeSpecName: "kube-api-access-cg7wd") pod "0e10dfb5-71c9-4b4c-a520-0845163d3b42" (UID: "0e10dfb5-71c9-4b4c-a520-0845163d3b42"). InnerVolumeSpecName "kube-api-access-cg7wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.600433 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0e10dfb5-71c9-4b4c-a520-0845163d3b42" (UID: "0e10dfb5-71c9-4b4c-a520-0845163d3b42"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.605939 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e10dfb5-71c9-4b4c-a520-0845163d3b42" (UID: "0e10dfb5-71c9-4b4c-a520-0845163d3b42"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.612045 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-config" (OuterVolumeSpecName: "config") pod "0e10dfb5-71c9-4b4c-a520-0845163d3b42" (UID: "0e10dfb5-71c9-4b4c-a520-0845163d3b42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.623688 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0e10dfb5-71c9-4b4c-a520-0845163d3b42" (UID: "0e10dfb5-71c9-4b4c-a520-0845163d3b42"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.654495 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.654525 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.654538 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg7wd\" (UniqueName: \"kubernetes.io/projected/0e10dfb5-71c9-4b4c-a520-0845163d3b42-kube-api-access-cg7wd\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.654546 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.654554 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e10dfb5-71c9-4b4c-a520-0845163d3b42-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:39 crc kubenswrapper[5035]: I1124 19:36:39.664668 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c58867b6c-btg8h"] Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.164128 5035 generic.go:334] "Generic (PLEG): container finished" podID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerID="25b5bc190054feab413a22a038a720c7556311acd08fc2c49bae6c89fc6a72a5" exitCode=0 Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.164187 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" event={"ID":"268a0d3d-f4b4-49c2-ae77-9ec588442031","Type":"ContainerDied","Data":"25b5bc190054feab413a22a038a720c7556311acd08fc2c49bae6c89fc6a72a5"} Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.164413 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" event={"ID":"268a0d3d-f4b4-49c2-ae77-9ec588442031","Type":"ContainerStarted","Data":"3f224e6f1e1a6a00662e3ad9be2d83de06207dbe22467572d25b88dd5839a201"} Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.166998 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.167030 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f95c456cf-dz6jb" event={"ID":"0e10dfb5-71c9-4b4c-a520-0845163d3b42","Type":"ContainerDied","Data":"7dc2a70ad99427b333b728505aa691b1d495c0b5fb1454e0015074b48b888652"} Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.167083 5035 scope.go:117] "RemoveContainer" containerID="93b09cd18ef0bac4192a0c4d0ae1a671fd2cfc5fd437e2653f4e0176ad86ee34" Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.168730 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" event={"ID":"82a1649c-9b91-4ed5-bff3-acddd2292b4f","Type":"ContainerStarted","Data":"a28a1ad6b71c735874f3eec6177bffc3109050a8d7da8bd5407c3ee0381078d6"} Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.195402 5035 scope.go:117] "RemoveContainer" containerID="2b5e3d5b19b7e67599e8bc8b7f2ccc8d2971262f25bbf5955762dbb2425679d9" Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.217079 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" podStartSLOduration=2.747214589 podStartE2EDuration="12.217055483s" podCreationTimestamp="2025-11-24 19:36:28 +0000 UTC" firstStartedPulling="2025-11-24 19:36:29.68633394 +0000 UTC m=+1328.208840217" lastFinishedPulling="2025-11-24 19:36:39.156174854 +0000 UTC m=+1337.678681111" observedRunningTime="2025-11-24 19:36:40.215948333 +0000 UTC m=+1338.738454610" watchObservedRunningTime="2025-11-24 19:36:40.217055483 +0000 UTC m=+1338.739561760" Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.244738 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f95c456cf-dz6jb"] Nov 24 19:36:40 crc kubenswrapper[5035]: I1124 19:36:40.255430 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f95c456cf-dz6jb"] Nov 24 19:36:41 crc kubenswrapper[5035]: I1124 19:36:41.182403 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" event={"ID":"268a0d3d-f4b4-49c2-ae77-9ec588442031","Type":"ContainerStarted","Data":"d7e1af7a46b48ded1939fb8cf0093593013ea80c00a3b9891de65e1fab8402d7"} Nov 24 19:36:41 crc kubenswrapper[5035]: I1124 19:36:41.184262 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:41 crc kubenswrapper[5035]: I1124 19:36:41.203847 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" podStartSLOduration=3.203831292 podStartE2EDuration="3.203831292s" podCreationTimestamp="2025-11-24 19:36:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:36:41.197982124 +0000 UTC m=+1339.720488391" watchObservedRunningTime="2025-11-24 19:36:41.203831292 +0000 UTC m=+1339.726337549" Nov 24 19:36:42 crc kubenswrapper[5035]: I1124 19:36:42.237049 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" path="/var/lib/kubelet/pods/0e10dfb5-71c9-4b4c-a520-0845163d3b42/volumes" Nov 24 19:36:45 crc kubenswrapper[5035]: I1124 19:36:45.233777 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:36:45 crc kubenswrapper[5035]: I1124 19:36:45.234045 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:36:48 crc kubenswrapper[5035]: I1124 19:36:48.591493 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 19:36:48 crc kubenswrapper[5035]: I1124 19:36:48.656035 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64b6dd64c5-mrbkx"] Nov 24 19:36:48 crc kubenswrapper[5035]: I1124 19:36:48.656581 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" podUID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerName="dnsmasq-dns" containerID="cri-o://5b2b0e6e3f5ceee94ac8b82d7f324af5305cce4649b182cf54ebef3ede9f6b02" gracePeriod=10 Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.256949 5035 generic.go:334] "Generic (PLEG): container finished" podID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerID="5b2b0e6e3f5ceee94ac8b82d7f324af5305cce4649b182cf54ebef3ede9f6b02" exitCode=0 Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.257007 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" event={"ID":"5426a0ef-fa48-4a37-a7b7-98601bb592a1","Type":"ContainerDied","Data":"5b2b0e6e3f5ceee94ac8b82d7f324af5305cce4649b182cf54ebef3ede9f6b02"} Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.628836 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.691895 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-sb\") pod \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.691950 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-nb\") pod \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.692062 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-config\") pod \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.692089 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-dns-svc\") pod \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.692888 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc9n5\" (UniqueName: \"kubernetes.io/projected/5426a0ef-fa48-4a37-a7b7-98601bb592a1-kube-api-access-zc9n5\") pod \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.692942 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-openstack-edpm-ipam\") pod \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\" (UID: \"5426a0ef-fa48-4a37-a7b7-98601bb592a1\") " Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.705742 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5426a0ef-fa48-4a37-a7b7-98601bb592a1-kube-api-access-zc9n5" (OuterVolumeSpecName: "kube-api-access-zc9n5") pod "5426a0ef-fa48-4a37-a7b7-98601bb592a1" (UID: "5426a0ef-fa48-4a37-a7b7-98601bb592a1"). InnerVolumeSpecName "kube-api-access-zc9n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.795303 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5426a0ef-fa48-4a37-a7b7-98601bb592a1" (UID: "5426a0ef-fa48-4a37-a7b7-98601bb592a1"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.795588 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc9n5\" (UniqueName: \"kubernetes.io/projected/5426a0ef-fa48-4a37-a7b7-98601bb592a1-kube-api-access-zc9n5\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.795612 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.799160 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5426a0ef-fa48-4a37-a7b7-98601bb592a1" (UID: "5426a0ef-fa48-4a37-a7b7-98601bb592a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.803701 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5426a0ef-fa48-4a37-a7b7-98601bb592a1" (UID: "5426a0ef-fa48-4a37-a7b7-98601bb592a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.805212 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5426a0ef-fa48-4a37-a7b7-98601bb592a1" (UID: "5426a0ef-fa48-4a37-a7b7-98601bb592a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.825911 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-config" (OuterVolumeSpecName: "config") pod "5426a0ef-fa48-4a37-a7b7-98601bb592a1" (UID: "5426a0ef-fa48-4a37-a7b7-98601bb592a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.896682 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.896709 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.896718 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-config\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:49 crc kubenswrapper[5035]: I1124 19:36:49.896727 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5426a0ef-fa48-4a37-a7b7-98601bb592a1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:50 crc kubenswrapper[5035]: I1124 19:36:50.270691 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" event={"ID":"5426a0ef-fa48-4a37-a7b7-98601bb592a1","Type":"ContainerDied","Data":"675b72838396d7fec753d17194ad4b64aaa8be7060317119391e6f1e7cd3f739"} Nov 24 19:36:50 crc kubenswrapper[5035]: I1124 19:36:50.270750 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b6dd64c5-mrbkx" Nov 24 19:36:50 crc kubenswrapper[5035]: I1124 19:36:50.270780 5035 scope.go:117] "RemoveContainer" containerID="5b2b0e6e3f5ceee94ac8b82d7f324af5305cce4649b182cf54ebef3ede9f6b02" Nov 24 19:36:50 crc kubenswrapper[5035]: I1124 19:36:50.296257 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64b6dd64c5-mrbkx"] Nov 24 19:36:50 crc kubenswrapper[5035]: I1124 19:36:50.301464 5035 scope.go:117] "RemoveContainer" containerID="c2e1cd285bfe278236bc95b13d43a8aeaa7477123afdd6e34d0e98da412beb22" Nov 24 19:36:50 crc kubenswrapper[5035]: I1124 19:36:50.306034 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64b6dd64c5-mrbkx"] Nov 24 19:36:52 crc kubenswrapper[5035]: I1124 19:36:52.210847 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" path="/var/lib/kubelet/pods/5426a0ef-fa48-4a37-a7b7-98601bb592a1/volumes" Nov 24 19:36:52 crc kubenswrapper[5035]: I1124 19:36:52.293865 5035 generic.go:334] "Generic (PLEG): container finished" podID="82a1649c-9b91-4ed5-bff3-acddd2292b4f" containerID="a28a1ad6b71c735874f3eec6177bffc3109050a8d7da8bd5407c3ee0381078d6" exitCode=0 Nov 24 19:36:52 crc kubenswrapper[5035]: I1124 19:36:52.293908 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" event={"ID":"82a1649c-9b91-4ed5-bff3-acddd2292b4f","Type":"ContainerDied","Data":"a28a1ad6b71c735874f3eec6177bffc3109050a8d7da8bd5407c3ee0381078d6"} Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.303994 5035 generic.go:334] "Generic (PLEG): container finished" podID="00618566-3b37-4b50-92d5-5f7dabf6a55b" containerID="11c83ac000c4cfea2e982277a67e6a9e0bf6fda4204328d4fb6d8bdf4ac2357c" exitCode=0 Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.304076 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00618566-3b37-4b50-92d5-5f7dabf6a55b","Type":"ContainerDied","Data":"11c83ac000c4cfea2e982277a67e6a9e0bf6fda4204328d4fb6d8bdf4ac2357c"} Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.709544 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.870661 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p84kg\" (UniqueName: \"kubernetes.io/projected/82a1649c-9b91-4ed5-bff3-acddd2292b4f-kube-api-access-p84kg\") pod \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.870752 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-ssh-key\") pod \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.870895 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-inventory\") pod \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.870925 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-repo-setup-combined-ca-bundle\") pod \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\" (UID: \"82a1649c-9b91-4ed5-bff3-acddd2292b4f\") " Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.877300 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "82a1649c-9b91-4ed5-bff3-acddd2292b4f" (UID: "82a1649c-9b91-4ed5-bff3-acddd2292b4f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.877974 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82a1649c-9b91-4ed5-bff3-acddd2292b4f-kube-api-access-p84kg" (OuterVolumeSpecName: "kube-api-access-p84kg") pod "82a1649c-9b91-4ed5-bff3-acddd2292b4f" (UID: "82a1649c-9b91-4ed5-bff3-acddd2292b4f"). InnerVolumeSpecName "kube-api-access-p84kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.906380 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-inventory" (OuterVolumeSpecName: "inventory") pod "82a1649c-9b91-4ed5-bff3-acddd2292b4f" (UID: "82a1649c-9b91-4ed5-bff3-acddd2292b4f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.918543 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "82a1649c-9b91-4ed5-bff3-acddd2292b4f" (UID: "82a1649c-9b91-4ed5-bff3-acddd2292b4f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.973855 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p84kg\" (UniqueName: \"kubernetes.io/projected/82a1649c-9b91-4ed5-bff3-acddd2292b4f-kube-api-access-p84kg\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.973914 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.973935 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:53 crc kubenswrapper[5035]: I1124 19:36:53.973950 5035 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a1649c-9b91-4ed5-bff3-acddd2292b4f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.321439 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00618566-3b37-4b50-92d5-5f7dabf6a55b","Type":"ContainerStarted","Data":"eccf3b8fe3c847b413c0122dbb08ae0ed65aee267a5e78f05328f5454e5a81f0"} Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.322227 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.323253 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" event={"ID":"82a1649c-9b91-4ed5-bff3-acddd2292b4f","Type":"ContainerDied","Data":"fb50b2681360f22c6f94d314b30405fdf2e853223676a4d6a6f011e8ecb89dcb"} Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.323287 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb50b2681360f22c6f94d314b30405fdf2e853223676a4d6a6f011e8ecb89dcb" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.323322 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.368219 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.368197009 podStartE2EDuration="37.368197009s" podCreationTimestamp="2025-11-24 19:36:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:36:54.351952628 +0000 UTC m=+1352.874458895" watchObservedRunningTime="2025-11-24 19:36:54.368197009 +0000 UTC m=+1352.890703266" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.476608 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw"] Nov 24 19:36:54 crc kubenswrapper[5035]: E1124 19:36:54.477061 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a1649c-9b91-4ed5-bff3-acddd2292b4f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477085 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a1649c-9b91-4ed5-bff3-acddd2292b4f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 19:36:54 crc kubenswrapper[5035]: E1124 19:36:54.477101 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerName="dnsmasq-dns" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477110 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerName="dnsmasq-dns" Nov 24 19:36:54 crc kubenswrapper[5035]: E1124 19:36:54.477130 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerName="dnsmasq-dns" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477138 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerName="dnsmasq-dns" Nov 24 19:36:54 crc kubenswrapper[5035]: E1124 19:36:54.477161 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerName="init" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477172 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerName="init" Nov 24 19:36:54 crc kubenswrapper[5035]: E1124 19:36:54.477202 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerName="init" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477212 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerName="init" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477535 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5426a0ef-fa48-4a37-a7b7-98601bb592a1" containerName="dnsmasq-dns" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477577 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e10dfb5-71c9-4b4c-a520-0845163d3b42" containerName="dnsmasq-dns" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.477602 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="82a1649c-9b91-4ed5-bff3-acddd2292b4f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.478573 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.482084 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.483224 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.483271 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.483583 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.488311 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw"] Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.586185 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqm5q\" (UniqueName: \"kubernetes.io/projected/5ee29809-634e-4c86-9605-5368bfc80b36-kube-api-access-hqm5q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.586377 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.586665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.586792 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.687976 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.688098 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.688145 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.688206 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqm5q\" (UniqueName: \"kubernetes.io/projected/5ee29809-634e-4c86-9605-5368bfc80b36-kube-api-access-hqm5q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.693118 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.693174 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.704221 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.709917 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqm5q\" (UniqueName: \"kubernetes.io/projected/5ee29809-634e-4c86-9605-5368bfc80b36-kube-api-access-hqm5q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:54 crc kubenswrapper[5035]: I1124 19:36:54.796369 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:36:55 crc kubenswrapper[5035]: I1124 19:36:55.288044 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw"] Nov 24 19:36:55 crc kubenswrapper[5035]: W1124 19:36:55.293718 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ee29809_634e_4c86_9605_5368bfc80b36.slice/crio-648e7fd4afe61661410ecc925718fb39f304a39a6c665bcf88bff1242605d566 WatchSource:0}: Error finding container 648e7fd4afe61661410ecc925718fb39f304a39a6c665bcf88bff1242605d566: Status 404 returned error can't find the container with id 648e7fd4afe61661410ecc925718fb39f304a39a6c665bcf88bff1242605d566 Nov 24 19:36:55 crc kubenswrapper[5035]: I1124 19:36:55.334796 5035 generic.go:334] "Generic (PLEG): container finished" podID="6d6f3cf9-b4c1-48e1-880b-8812b7f4421c" containerID="c91cab0b50981d6d3cf9c6d236923303448c342ae9118b3a328fc3fb6b80ae19" exitCode=0 Nov 24 19:36:55 crc kubenswrapper[5035]: I1124 19:36:55.334898 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c","Type":"ContainerDied","Data":"c91cab0b50981d6d3cf9c6d236923303448c342ae9118b3a328fc3fb6b80ae19"} Nov 24 19:36:55 crc kubenswrapper[5035]: I1124 19:36:55.338036 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" event={"ID":"5ee29809-634e-4c86-9605-5368bfc80b36","Type":"ContainerStarted","Data":"648e7fd4afe61661410ecc925718fb39f304a39a6c665bcf88bff1242605d566"} Nov 24 19:36:56 crc kubenswrapper[5035]: I1124 19:36:56.349669 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" event={"ID":"5ee29809-634e-4c86-9605-5368bfc80b36","Type":"ContainerStarted","Data":"7107bfbed7b0fc114c81497ac87bcb2ddecd6b064f06f2d6a4d76b0601f54215"} Nov 24 19:36:56 crc kubenswrapper[5035]: I1124 19:36:56.363822 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d6f3cf9-b4c1-48e1-880b-8812b7f4421c","Type":"ContainerStarted","Data":"c3b317393483d4f02715c8d7cb725aaefb57cb515d188ec0076d31ce6a08590f"} Nov 24 19:36:56 crc kubenswrapper[5035]: I1124 19:36:56.364833 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:36:56 crc kubenswrapper[5035]: I1124 19:36:56.375229 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" podStartSLOduration=1.820771185 podStartE2EDuration="2.375208965s" podCreationTimestamp="2025-11-24 19:36:54 +0000 UTC" firstStartedPulling="2025-11-24 19:36:55.296231955 +0000 UTC m=+1353.818738222" lastFinishedPulling="2025-11-24 19:36:55.850669735 +0000 UTC m=+1354.373176002" observedRunningTime="2025-11-24 19:36:56.373833538 +0000 UTC m=+1354.896339795" watchObservedRunningTime="2025-11-24 19:36:56.375208965 +0000 UTC m=+1354.897715222" Nov 24 19:36:56 crc kubenswrapper[5035]: I1124 19:36:56.411941 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.411911091 podStartE2EDuration="38.411911091s" podCreationTimestamp="2025-11-24 19:36:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 19:36:56.411713655 +0000 UTC m=+1354.934219922" watchObservedRunningTime="2025-11-24 19:36:56.411911091 +0000 UTC m=+1354.934417348" Nov 24 19:37:08 crc kubenswrapper[5035]: I1124 19:37:08.639517 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 19:37:09 crc kubenswrapper[5035]: I1124 19:37:09.372478 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.234971 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.235765 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.235840 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.237019 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e56edadec21ff9db609cb1eb46281d662ed2c26ae0bd15afd911d69fb3661e1"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.237139 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://5e56edadec21ff9db609cb1eb46281d662ed2c26ae0bd15afd911d69fb3661e1" gracePeriod=600 Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.563724 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="5e56edadec21ff9db609cb1eb46281d662ed2c26ae0bd15afd911d69fb3661e1" exitCode=0 Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.563918 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"5e56edadec21ff9db609cb1eb46281d662ed2c26ae0bd15afd911d69fb3661e1"} Nov 24 19:37:15 crc kubenswrapper[5035]: I1124 19:37:15.564162 5035 scope.go:117] "RemoveContainer" containerID="d2a1e9879f1117f6209b6479ae153039678ddcd07bf4e0e5400eeea2f3c473b0" Nov 24 19:37:16 crc kubenswrapper[5035]: I1124 19:37:16.579473 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199"} Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.081959 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2z6cz"] Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.086622 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.140774 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2z6cz"] Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.271119 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cq6g\" (UniqueName: \"kubernetes.io/projected/52a63a97-8196-4344-bec6-df0c09b57642-kube-api-access-8cq6g\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.271462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a63a97-8196-4344-bec6-df0c09b57642-catalog-content\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.271502 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a63a97-8196-4344-bec6-df0c09b57642-utilities\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.373438 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cq6g\" (UniqueName: \"kubernetes.io/projected/52a63a97-8196-4344-bec6-df0c09b57642-kube-api-access-8cq6g\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.373542 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a63a97-8196-4344-bec6-df0c09b57642-catalog-content\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.373572 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a63a97-8196-4344-bec6-df0c09b57642-utilities\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.373958 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a63a97-8196-4344-bec6-df0c09b57642-catalog-content\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.374362 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a63a97-8196-4344-bec6-df0c09b57642-utilities\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.396088 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cq6g\" (UniqueName: \"kubernetes.io/projected/52a63a97-8196-4344-bec6-df0c09b57642-kube-api-access-8cq6g\") pod \"redhat-operators-2z6cz\" (UID: \"52a63a97-8196-4344-bec6-df0c09b57642\") " pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.458796 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:09 crc kubenswrapper[5035]: I1124 19:38:09.895068 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2z6cz"] Nov 24 19:38:10 crc kubenswrapper[5035]: I1124 19:38:10.150619 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2z6cz" event={"ID":"52a63a97-8196-4344-bec6-df0c09b57642","Type":"ContainerStarted","Data":"42f21051b4800ed0a697f73a4c23da59be7c07be092e1d801633648dd1f41434"} Nov 24 19:38:10 crc kubenswrapper[5035]: I1124 19:38:10.150911 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2z6cz" event={"ID":"52a63a97-8196-4344-bec6-df0c09b57642","Type":"ContainerStarted","Data":"8d76d48ec4d0e4cc1af26f8816b2e858721057d3c1c5a3fa278998ad0c88d5c4"} Nov 24 19:38:11 crc kubenswrapper[5035]: I1124 19:38:11.163745 5035 generic.go:334] "Generic (PLEG): container finished" podID="52a63a97-8196-4344-bec6-df0c09b57642" containerID="42f21051b4800ed0a697f73a4c23da59be7c07be092e1d801633648dd1f41434" exitCode=0 Nov 24 19:38:11 crc kubenswrapper[5035]: I1124 19:38:11.163817 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2z6cz" event={"ID":"52a63a97-8196-4344-bec6-df0c09b57642","Type":"ContainerDied","Data":"42f21051b4800ed0a697f73a4c23da59be7c07be092e1d801633648dd1f41434"} Nov 24 19:38:19 crc kubenswrapper[5035]: I1124 19:38:19.243080 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2z6cz" event={"ID":"52a63a97-8196-4344-bec6-df0c09b57642","Type":"ContainerStarted","Data":"69862c93ad14b24985c46c3c793c7425a83aa8f7dcc715788593f4fed174e770"} Nov 24 19:38:23 crc kubenswrapper[5035]: I1124 19:38:23.287725 5035 generic.go:334] "Generic (PLEG): container finished" podID="52a63a97-8196-4344-bec6-df0c09b57642" containerID="69862c93ad14b24985c46c3c793c7425a83aa8f7dcc715788593f4fed174e770" exitCode=0 Nov 24 19:38:23 crc kubenswrapper[5035]: I1124 19:38:23.287948 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2z6cz" event={"ID":"52a63a97-8196-4344-bec6-df0c09b57642","Type":"ContainerDied","Data":"69862c93ad14b24985c46c3c793c7425a83aa8f7dcc715788593f4fed174e770"} Nov 24 19:38:23 crc kubenswrapper[5035]: I1124 19:38:23.967571 5035 scope.go:117] "RemoveContainer" containerID="c39bf0115ddd19cdb161ec48d1442aa37d41032ad0df9379f64a096d3cdded02" Nov 24 19:38:24 crc kubenswrapper[5035]: I1124 19:38:24.300398 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2z6cz" event={"ID":"52a63a97-8196-4344-bec6-df0c09b57642","Type":"ContainerStarted","Data":"294f446a52e7546bc61ad688d2b358cb728f8980b1be606631f89280951cebf3"} Nov 24 19:38:24 crc kubenswrapper[5035]: I1124 19:38:24.323536 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2z6cz" podStartSLOduration=1.615523128 podStartE2EDuration="15.323517404s" podCreationTimestamp="2025-11-24 19:38:09 +0000 UTC" firstStartedPulling="2025-11-24 19:38:10.151886333 +0000 UTC m=+1428.674392590" lastFinishedPulling="2025-11-24 19:38:23.859880609 +0000 UTC m=+1442.382386866" observedRunningTime="2025-11-24 19:38:24.319972499 +0000 UTC m=+1442.842478776" watchObservedRunningTime="2025-11-24 19:38:24.323517404 +0000 UTC m=+1442.846023661" Nov 24 19:38:29 crc kubenswrapper[5035]: I1124 19:38:29.459474 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:29 crc kubenswrapper[5035]: I1124 19:38:29.460330 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:30 crc kubenswrapper[5035]: I1124 19:38:30.517926 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2z6cz" podUID="52a63a97-8196-4344-bec6-df0c09b57642" containerName="registry-server" probeResult="failure" output=< Nov 24 19:38:30 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 19:38:30 crc kubenswrapper[5035]: > Nov 24 19:38:36 crc kubenswrapper[5035]: I1124 19:38:36.796335 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hx92n"] Nov 24 19:38:36 crc kubenswrapper[5035]: I1124 19:38:36.798619 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:36 crc kubenswrapper[5035]: I1124 19:38:36.819771 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hx92n"] Nov 24 19:38:36 crc kubenswrapper[5035]: I1124 19:38:36.947407 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-utilities\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:36 crc kubenswrapper[5035]: I1124 19:38:36.947585 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-645hp\" (UniqueName: \"kubernetes.io/projected/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-kube-api-access-645hp\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:36 crc kubenswrapper[5035]: I1124 19:38:36.947707 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-catalog-content\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.049022 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-catalog-content\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.049165 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-utilities\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.049205 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-645hp\" (UniqueName: \"kubernetes.io/projected/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-kube-api-access-645hp\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.049741 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-catalog-content\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.049830 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-utilities\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.082210 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-645hp\" (UniqueName: \"kubernetes.io/projected/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-kube-api-access-645hp\") pod \"certified-operators-hx92n\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.137526 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:37 crc kubenswrapper[5035]: I1124 19:38:37.644433 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hx92n"] Nov 24 19:38:39 crc kubenswrapper[5035]: I1124 19:38:39.100253 5035 generic.go:334] "Generic (PLEG): container finished" podID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerID="20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c" exitCode=0 Nov 24 19:38:39 crc kubenswrapper[5035]: I1124 19:38:39.100343 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx92n" event={"ID":"7ab3af7e-e65a-4a94-a05f-8451fd8cac53","Type":"ContainerDied","Data":"20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c"} Nov 24 19:38:39 crc kubenswrapper[5035]: I1124 19:38:39.101067 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx92n" event={"ID":"7ab3af7e-e65a-4a94-a05f-8451fd8cac53","Type":"ContainerStarted","Data":"dae52d16b376d92ab430f701afdcd63282d3cd42689dc4153ca94d1c6f045827"} Nov 24 19:38:39 crc kubenswrapper[5035]: I1124 19:38:39.513932 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:39 crc kubenswrapper[5035]: I1124 19:38:39.556909 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2z6cz" Nov 24 19:38:40 crc kubenswrapper[5035]: I1124 19:38:40.112364 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx92n" event={"ID":"7ab3af7e-e65a-4a94-a05f-8451fd8cac53","Type":"ContainerStarted","Data":"e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4"} Nov 24 19:38:40 crc kubenswrapper[5035]: I1124 19:38:40.607823 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2z6cz"] Nov 24 19:38:40 crc kubenswrapper[5035]: I1124 19:38:40.967581 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s9kng"] Nov 24 19:38:40 crc kubenswrapper[5035]: I1124 19:38:40.968116 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s9kng" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="registry-server" containerID="cri-o://e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5" gracePeriod=2 Nov 24 19:38:41 crc kubenswrapper[5035]: I1124 19:38:41.130387 5035 generic.go:334] "Generic (PLEG): container finished" podID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerID="e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4" exitCode=0 Nov 24 19:38:41 crc kubenswrapper[5035]: I1124 19:38:41.130621 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx92n" event={"ID":"7ab3af7e-e65a-4a94-a05f-8451fd8cac53","Type":"ContainerDied","Data":"e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4"} Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.028233 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.141019 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx92n" event={"ID":"7ab3af7e-e65a-4a94-a05f-8451fd8cac53","Type":"ContainerStarted","Data":"159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc"} Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.144376 5035 generic.go:334] "Generic (PLEG): container finished" podID="62284293-00ac-4dd5-8e14-ecebe023008c" containerID="e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5" exitCode=0 Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.144409 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9kng" event={"ID":"62284293-00ac-4dd5-8e14-ecebe023008c","Type":"ContainerDied","Data":"e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5"} Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.144432 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s9kng" event={"ID":"62284293-00ac-4dd5-8e14-ecebe023008c","Type":"ContainerDied","Data":"161a013d6fbd24dbdd60ebaca4e098b641e3c4b1b957f1951d2cc32acf2b9479"} Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.144438 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s9kng" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.144448 5035 scope.go:117] "RemoveContainer" containerID="e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.168599 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hx92n" podStartSLOduration=3.740857905 podStartE2EDuration="6.168581659s" podCreationTimestamp="2025-11-24 19:38:36 +0000 UTC" firstStartedPulling="2025-11-24 19:38:39.102998383 +0000 UTC m=+1457.625504660" lastFinishedPulling="2025-11-24 19:38:41.530722157 +0000 UTC m=+1460.053228414" observedRunningTime="2025-11-24 19:38:42.158034783 +0000 UTC m=+1460.680541040" watchObservedRunningTime="2025-11-24 19:38:42.168581659 +0000 UTC m=+1460.691087916" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.169482 5035 scope.go:117] "RemoveContainer" containerID="2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.175674 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-utilities\") pod \"62284293-00ac-4dd5-8e14-ecebe023008c\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.175781 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-catalog-content\") pod \"62284293-00ac-4dd5-8e14-ecebe023008c\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.175945 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpnt6\" (UniqueName: \"kubernetes.io/projected/62284293-00ac-4dd5-8e14-ecebe023008c-kube-api-access-kpnt6\") pod \"62284293-00ac-4dd5-8e14-ecebe023008c\" (UID: \"62284293-00ac-4dd5-8e14-ecebe023008c\") " Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.177378 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-utilities" (OuterVolumeSpecName: "utilities") pod "62284293-00ac-4dd5-8e14-ecebe023008c" (UID: "62284293-00ac-4dd5-8e14-ecebe023008c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.186545 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62284293-00ac-4dd5-8e14-ecebe023008c-kube-api-access-kpnt6" (OuterVolumeSpecName: "kube-api-access-kpnt6") pod "62284293-00ac-4dd5-8e14-ecebe023008c" (UID: "62284293-00ac-4dd5-8e14-ecebe023008c"). InnerVolumeSpecName "kube-api-access-kpnt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.199229 5035 scope.go:117] "RemoveContainer" containerID="b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.277608 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpnt6\" (UniqueName: \"kubernetes.io/projected/62284293-00ac-4dd5-8e14-ecebe023008c-kube-api-access-kpnt6\") on node \"crc\" DevicePath \"\"" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.277638 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.282176 5035 scope.go:117] "RemoveContainer" containerID="e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5" Nov 24 19:38:42 crc kubenswrapper[5035]: E1124 19:38:42.283624 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5\": container with ID starting with e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5 not found: ID does not exist" containerID="e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.283663 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5"} err="failed to get container status \"e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5\": rpc error: code = NotFound desc = could not find container \"e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5\": container with ID starting with e6d623def77b1322a6c9f5c8045fbbdc20a8d2cf8b2a3124711fe8f8bd4b0da5 not found: ID does not exist" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.283688 5035 scope.go:117] "RemoveContainer" containerID="2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b" Nov 24 19:38:42 crc kubenswrapper[5035]: E1124 19:38:42.285843 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b\": container with ID starting with 2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b not found: ID does not exist" containerID="2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.285868 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b"} err="failed to get container status \"2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b\": rpc error: code = NotFound desc = could not find container \"2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b\": container with ID starting with 2dc3293d13b67b026d692d2337ccd34e5ad18ed181a833b802d92ddbfe945e6b not found: ID does not exist" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.285883 5035 scope.go:117] "RemoveContainer" containerID="b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12" Nov 24 19:38:42 crc kubenswrapper[5035]: E1124 19:38:42.288936 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12\": container with ID starting with b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12 not found: ID does not exist" containerID="b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.288964 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12"} err="failed to get container status \"b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12\": rpc error: code = NotFound desc = could not find container \"b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12\": container with ID starting with b9adfb21f7339bbab3f3767df1aee5fc52862e8cadbcab3d1b3c98decb5c3d12 not found: ID does not exist" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.296846 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62284293-00ac-4dd5-8e14-ecebe023008c" (UID: "62284293-00ac-4dd5-8e14-ecebe023008c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.379565 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62284293-00ac-4dd5-8e14-ecebe023008c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.476198 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s9kng"] Nov 24 19:38:42 crc kubenswrapper[5035]: I1124 19:38:42.484217 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s9kng"] Nov 24 19:38:44 crc kubenswrapper[5035]: I1124 19:38:44.211313 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" path="/var/lib/kubelet/pods/62284293-00ac-4dd5-8e14-ecebe023008c/volumes" Nov 24 19:38:47 crc kubenswrapper[5035]: I1124 19:38:47.138344 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:47 crc kubenswrapper[5035]: I1124 19:38:47.139894 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:47 crc kubenswrapper[5035]: I1124 19:38:47.198088 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:47 crc kubenswrapper[5035]: I1124 19:38:47.256063 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:48 crc kubenswrapper[5035]: I1124 19:38:48.367898 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hx92n"] Nov 24 19:38:50 crc kubenswrapper[5035]: I1124 19:38:50.232152 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hx92n" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="registry-server" containerID="cri-o://159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc" gracePeriod=2 Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.221752 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.252844 5035 generic.go:334] "Generic (PLEG): container finished" podID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerID="159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc" exitCode=0 Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.253391 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx92n" event={"ID":"7ab3af7e-e65a-4a94-a05f-8451fd8cac53","Type":"ContainerDied","Data":"159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc"} Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.253536 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx92n" event={"ID":"7ab3af7e-e65a-4a94-a05f-8451fd8cac53","Type":"ContainerDied","Data":"dae52d16b376d92ab430f701afdcd63282d3cd42689dc4153ca94d1c6f045827"} Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.253622 5035 scope.go:117] "RemoveContainer" containerID="159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.253828 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx92n" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.280093 5035 scope.go:117] "RemoveContainer" containerID="e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.309052 5035 scope.go:117] "RemoveContainer" containerID="20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.340762 5035 scope.go:117] "RemoveContainer" containerID="159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.342420 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-645hp\" (UniqueName: \"kubernetes.io/projected/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-kube-api-access-645hp\") pod \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.342655 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-catalog-content\") pod \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.342766 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-utilities\") pod \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\" (UID: \"7ab3af7e-e65a-4a94-a05f-8451fd8cac53\") " Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.344510 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-utilities" (OuterVolumeSpecName: "utilities") pod "7ab3af7e-e65a-4a94-a05f-8451fd8cac53" (UID: "7ab3af7e-e65a-4a94-a05f-8451fd8cac53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:38:51 crc kubenswrapper[5035]: E1124 19:38:51.344905 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc\": container with ID starting with 159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc not found: ID does not exist" containerID="159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.344965 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc"} err="failed to get container status \"159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc\": rpc error: code = NotFound desc = could not find container \"159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc\": container with ID starting with 159997585f6d08c560ae07de34a9760acac432727185a64662fce647f9905adc not found: ID does not exist" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.345008 5035 scope.go:117] "RemoveContainer" containerID="e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4" Nov 24 19:38:51 crc kubenswrapper[5035]: E1124 19:38:51.345584 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4\": container with ID starting with e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4 not found: ID does not exist" containerID="e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.345628 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4"} err="failed to get container status \"e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4\": rpc error: code = NotFound desc = could not find container \"e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4\": container with ID starting with e014e551677e44116289c3bd46038365bb19d38b450889029116c205cd33eef4 not found: ID does not exist" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.345654 5035 scope.go:117] "RemoveContainer" containerID="20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c" Nov 24 19:38:51 crc kubenswrapper[5035]: E1124 19:38:51.346198 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c\": container with ID starting with 20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c not found: ID does not exist" containerID="20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.346260 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c"} err="failed to get container status \"20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c\": rpc error: code = NotFound desc = could not find container \"20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c\": container with ID starting with 20df3698536a1d8b705df41864c9fd924e1245912c9e72526fd25b306a801c2c not found: ID does not exist" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.349495 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-kube-api-access-645hp" (OuterVolumeSpecName: "kube-api-access-645hp") pod "7ab3af7e-e65a-4a94-a05f-8451fd8cac53" (UID: "7ab3af7e-e65a-4a94-a05f-8451fd8cac53"). InnerVolumeSpecName "kube-api-access-645hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.394188 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ab3af7e-e65a-4a94-a05f-8451fd8cac53" (UID: "7ab3af7e-e65a-4a94-a05f-8451fd8cac53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.445190 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-645hp\" (UniqueName: \"kubernetes.io/projected/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-kube-api-access-645hp\") on node \"crc\" DevicePath \"\"" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.445447 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.445511 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ab3af7e-e65a-4a94-a05f-8451fd8cac53-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.597949 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hx92n"] Nov 24 19:38:51 crc kubenswrapper[5035]: I1124 19:38:51.606130 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hx92n"] Nov 24 19:38:52 crc kubenswrapper[5035]: I1124 19:38:52.218097 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" path="/var/lib/kubelet/pods/7ab3af7e-e65a-4a94-a05f-8451fd8cac53/volumes" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.753971 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5tgtg"] Nov 24 19:39:00 crc kubenswrapper[5035]: E1124 19:39:00.755007 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="registry-server" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755020 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="registry-server" Nov 24 19:39:00 crc kubenswrapper[5035]: E1124 19:39:00.755041 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="extract-utilities" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755049 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="extract-utilities" Nov 24 19:39:00 crc kubenswrapper[5035]: E1124 19:39:00.755073 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="extract-content" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755080 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="extract-content" Nov 24 19:39:00 crc kubenswrapper[5035]: E1124 19:39:00.755114 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="registry-server" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755120 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="registry-server" Nov 24 19:39:00 crc kubenswrapper[5035]: E1124 19:39:00.755131 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="extract-utilities" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755137 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="extract-utilities" Nov 24 19:39:00 crc kubenswrapper[5035]: E1124 19:39:00.755159 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="extract-content" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755164 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="extract-content" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755456 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ab3af7e-e65a-4a94-a05f-8451fd8cac53" containerName="registry-server" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.755495 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="62284293-00ac-4dd5-8e14-ecebe023008c" containerName="registry-server" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.757409 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.781508 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tgtg"] Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.941147 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-catalog-content\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.941379 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-utilities\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:00 crc kubenswrapper[5035]: I1124 19:39:00.941409 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knpgh\" (UniqueName: \"kubernetes.io/projected/22efa5a8-e97f-4702-9a47-626b05cdb13a-kube-api-access-knpgh\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.042725 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-catalog-content\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.043085 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-utilities\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.043176 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knpgh\" (UniqueName: \"kubernetes.io/projected/22efa5a8-e97f-4702-9a47-626b05cdb13a-kube-api-access-knpgh\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.043726 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-catalog-content\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.044016 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-utilities\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.078582 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knpgh\" (UniqueName: \"kubernetes.io/projected/22efa5a8-e97f-4702-9a47-626b05cdb13a-kube-api-access-knpgh\") pod \"redhat-marketplace-5tgtg\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.098196 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:01 crc kubenswrapper[5035]: I1124 19:39:01.554339 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tgtg"] Nov 24 19:39:02 crc kubenswrapper[5035]: I1124 19:39:02.401529 5035 generic.go:334] "Generic (PLEG): container finished" podID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerID="1f36c22ea2fcdb76e4f77e39a76ddda86e0b2ca386add147433a833e2949e68e" exitCode=0 Nov 24 19:39:02 crc kubenswrapper[5035]: I1124 19:39:02.401652 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tgtg" event={"ID":"22efa5a8-e97f-4702-9a47-626b05cdb13a","Type":"ContainerDied","Data":"1f36c22ea2fcdb76e4f77e39a76ddda86e0b2ca386add147433a833e2949e68e"} Nov 24 19:39:02 crc kubenswrapper[5035]: I1124 19:39:02.401789 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tgtg" event={"ID":"22efa5a8-e97f-4702-9a47-626b05cdb13a","Type":"ContainerStarted","Data":"86eaddd9f9ce4ddf50dc9dffd133696226212bf85226c71009188b1a5aa3d563"} Nov 24 19:39:03 crc kubenswrapper[5035]: I1124 19:39:03.414182 5035 generic.go:334] "Generic (PLEG): container finished" podID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerID="e56bf8c1e6cddfddbed2480e5692e72a825a250a8e1cf9e685bd2a77bd373916" exitCode=0 Nov 24 19:39:03 crc kubenswrapper[5035]: I1124 19:39:03.414238 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tgtg" event={"ID":"22efa5a8-e97f-4702-9a47-626b05cdb13a","Type":"ContainerDied","Data":"e56bf8c1e6cddfddbed2480e5692e72a825a250a8e1cf9e685bd2a77bd373916"} Nov 24 19:39:04 crc kubenswrapper[5035]: I1124 19:39:04.433832 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tgtg" event={"ID":"22efa5a8-e97f-4702-9a47-626b05cdb13a","Type":"ContainerStarted","Data":"b5223d0a74f3ad3936dff3f303b7933a624ec0175ef8185d0bd50636ec2e1723"} Nov 24 19:39:04 crc kubenswrapper[5035]: I1124 19:39:04.459891 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5tgtg" podStartSLOduration=3.019053355 podStartE2EDuration="4.459868107s" podCreationTimestamp="2025-11-24 19:39:00 +0000 UTC" firstStartedPulling="2025-11-24 19:39:02.40352479 +0000 UTC m=+1480.926031057" lastFinishedPulling="2025-11-24 19:39:03.844339552 +0000 UTC m=+1482.366845809" observedRunningTime="2025-11-24 19:39:04.451627413 +0000 UTC m=+1482.974133710" watchObservedRunningTime="2025-11-24 19:39:04.459868107 +0000 UTC m=+1482.982374394" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.494034 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rkw2f"] Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.496717 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.524268 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rkw2f"] Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.628805 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-catalog-content\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.628867 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmp6m\" (UniqueName: \"kubernetes.io/projected/fef91184-d485-4f47-abc8-7b7aac22d9d1-kube-api-access-xmp6m\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.629129 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-utilities\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.731223 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmp6m\" (UniqueName: \"kubernetes.io/projected/fef91184-d485-4f47-abc8-7b7aac22d9d1-kube-api-access-xmp6m\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.731356 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-utilities\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.731501 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-catalog-content\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.731948 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-catalog-content\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.732217 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-utilities\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.757800 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmp6m\" (UniqueName: \"kubernetes.io/projected/fef91184-d485-4f47-abc8-7b7aac22d9d1-kube-api-access-xmp6m\") pod \"community-operators-rkw2f\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:05 crc kubenswrapper[5035]: I1124 19:39:05.820103 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:06 crc kubenswrapper[5035]: I1124 19:39:06.409691 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rkw2f"] Nov 24 19:39:06 crc kubenswrapper[5035]: I1124 19:39:06.460622 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkw2f" event={"ID":"fef91184-d485-4f47-abc8-7b7aac22d9d1","Type":"ContainerStarted","Data":"e9adc99e288d8229bc09d0a36370f27e3da0dae0e4b599bcc96cfcb50a49d3bf"} Nov 24 19:39:07 crc kubenswrapper[5035]: I1124 19:39:07.477331 5035 generic.go:334] "Generic (PLEG): container finished" podID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerID="a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2" exitCode=0 Nov 24 19:39:07 crc kubenswrapper[5035]: I1124 19:39:07.477409 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkw2f" event={"ID":"fef91184-d485-4f47-abc8-7b7aac22d9d1","Type":"ContainerDied","Data":"a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2"} Nov 24 19:39:09 crc kubenswrapper[5035]: I1124 19:39:09.506870 5035 generic.go:334] "Generic (PLEG): container finished" podID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerID="a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76" exitCode=0 Nov 24 19:39:09 crc kubenswrapper[5035]: I1124 19:39:09.506947 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkw2f" event={"ID":"fef91184-d485-4f47-abc8-7b7aac22d9d1","Type":"ContainerDied","Data":"a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76"} Nov 24 19:39:10 crc kubenswrapper[5035]: I1124 19:39:10.519756 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkw2f" event={"ID":"fef91184-d485-4f47-abc8-7b7aac22d9d1","Type":"ContainerStarted","Data":"6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c"} Nov 24 19:39:10 crc kubenswrapper[5035]: I1124 19:39:10.543259 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rkw2f" podStartSLOduration=2.974124262 podStartE2EDuration="5.54323445s" podCreationTimestamp="2025-11-24 19:39:05 +0000 UTC" firstStartedPulling="2025-11-24 19:39:07.482112215 +0000 UTC m=+1486.004618512" lastFinishedPulling="2025-11-24 19:39:10.051222403 +0000 UTC m=+1488.573728700" observedRunningTime="2025-11-24 19:39:10.538192082 +0000 UTC m=+1489.060698359" watchObservedRunningTime="2025-11-24 19:39:10.54323445 +0000 UTC m=+1489.065740717" Nov 24 19:39:11 crc kubenswrapper[5035]: I1124 19:39:11.098866 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:11 crc kubenswrapper[5035]: I1124 19:39:11.099227 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:11 crc kubenswrapper[5035]: I1124 19:39:11.162368 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:11 crc kubenswrapper[5035]: I1124 19:39:11.589944 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:12 crc kubenswrapper[5035]: I1124 19:39:12.281771 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tgtg"] Nov 24 19:39:13 crc kubenswrapper[5035]: I1124 19:39:13.547031 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5tgtg" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="registry-server" containerID="cri-o://b5223d0a74f3ad3936dff3f303b7933a624ec0175ef8185d0bd50636ec2e1723" gracePeriod=2 Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.560107 5035 generic.go:334] "Generic (PLEG): container finished" podID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerID="b5223d0a74f3ad3936dff3f303b7933a624ec0175ef8185d0bd50636ec2e1723" exitCode=0 Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.560157 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tgtg" event={"ID":"22efa5a8-e97f-4702-9a47-626b05cdb13a","Type":"ContainerDied","Data":"b5223d0a74f3ad3936dff3f303b7933a624ec0175ef8185d0bd50636ec2e1723"} Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.560393 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tgtg" event={"ID":"22efa5a8-e97f-4702-9a47-626b05cdb13a","Type":"ContainerDied","Data":"86eaddd9f9ce4ddf50dc9dffd133696226212bf85226c71009188b1a5aa3d563"} Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.560409 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86eaddd9f9ce4ddf50dc9dffd133696226212bf85226c71009188b1a5aa3d563" Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.578654 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.729879 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-catalog-content\") pod \"22efa5a8-e97f-4702-9a47-626b05cdb13a\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.730076 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knpgh\" (UniqueName: \"kubernetes.io/projected/22efa5a8-e97f-4702-9a47-626b05cdb13a-kube-api-access-knpgh\") pod \"22efa5a8-e97f-4702-9a47-626b05cdb13a\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.730145 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-utilities\") pod \"22efa5a8-e97f-4702-9a47-626b05cdb13a\" (UID: \"22efa5a8-e97f-4702-9a47-626b05cdb13a\") " Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.731236 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-utilities" (OuterVolumeSpecName: "utilities") pod "22efa5a8-e97f-4702-9a47-626b05cdb13a" (UID: "22efa5a8-e97f-4702-9a47-626b05cdb13a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.740575 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22efa5a8-e97f-4702-9a47-626b05cdb13a-kube-api-access-knpgh" (OuterVolumeSpecName: "kube-api-access-knpgh") pod "22efa5a8-e97f-4702-9a47-626b05cdb13a" (UID: "22efa5a8-e97f-4702-9a47-626b05cdb13a"). InnerVolumeSpecName "kube-api-access-knpgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.745697 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22efa5a8-e97f-4702-9a47-626b05cdb13a" (UID: "22efa5a8-e97f-4702-9a47-626b05cdb13a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.832145 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.832177 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22efa5a8-e97f-4702-9a47-626b05cdb13a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:39:14 crc kubenswrapper[5035]: I1124 19:39:14.832189 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knpgh\" (UniqueName: \"kubernetes.io/projected/22efa5a8-e97f-4702-9a47-626b05cdb13a-kube-api-access-knpgh\") on node \"crc\" DevicePath \"\"" Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.234610 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.234670 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.573001 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tgtg" Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.622247 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tgtg"] Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.631596 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tgtg"] Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.820997 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.821358 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:15 crc kubenswrapper[5035]: I1124 19:39:15.876781 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:16 crc kubenswrapper[5035]: I1124 19:39:16.212082 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" path="/var/lib/kubelet/pods/22efa5a8-e97f-4702-9a47-626b05cdb13a/volumes" Nov 24 19:39:16 crc kubenswrapper[5035]: I1124 19:39:16.692544 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:18 crc kubenswrapper[5035]: I1124 19:39:18.286120 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rkw2f"] Nov 24 19:39:18 crc kubenswrapper[5035]: I1124 19:39:18.627567 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rkw2f" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="registry-server" containerID="cri-o://6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c" gracePeriod=2 Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.085399 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.233324 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-utilities\") pod \"fef91184-d485-4f47-abc8-7b7aac22d9d1\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.233492 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-catalog-content\") pod \"fef91184-d485-4f47-abc8-7b7aac22d9d1\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.233746 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmp6m\" (UniqueName: \"kubernetes.io/projected/fef91184-d485-4f47-abc8-7b7aac22d9d1-kube-api-access-xmp6m\") pod \"fef91184-d485-4f47-abc8-7b7aac22d9d1\" (UID: \"fef91184-d485-4f47-abc8-7b7aac22d9d1\") " Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.234521 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-utilities" (OuterVolumeSpecName: "utilities") pod "fef91184-d485-4f47-abc8-7b7aac22d9d1" (UID: "fef91184-d485-4f47-abc8-7b7aac22d9d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.234837 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.241810 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef91184-d485-4f47-abc8-7b7aac22d9d1-kube-api-access-xmp6m" (OuterVolumeSpecName: "kube-api-access-xmp6m") pod "fef91184-d485-4f47-abc8-7b7aac22d9d1" (UID: "fef91184-d485-4f47-abc8-7b7aac22d9d1"). InnerVolumeSpecName "kube-api-access-xmp6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.311334 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fef91184-d485-4f47-abc8-7b7aac22d9d1" (UID: "fef91184-d485-4f47-abc8-7b7aac22d9d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.336913 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef91184-d485-4f47-abc8-7b7aac22d9d1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.336963 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmp6m\" (UniqueName: \"kubernetes.io/projected/fef91184-d485-4f47-abc8-7b7aac22d9d1-kube-api-access-xmp6m\") on node \"crc\" DevicePath \"\"" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.641134 5035 generic.go:334] "Generic (PLEG): container finished" podID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerID="6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c" exitCode=0 Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.641218 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkw2f" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.641221 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkw2f" event={"ID":"fef91184-d485-4f47-abc8-7b7aac22d9d1","Type":"ContainerDied","Data":"6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c"} Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.641742 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkw2f" event={"ID":"fef91184-d485-4f47-abc8-7b7aac22d9d1","Type":"ContainerDied","Data":"e9adc99e288d8229bc09d0a36370f27e3da0dae0e4b599bcc96cfcb50a49d3bf"} Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.641788 5035 scope.go:117] "RemoveContainer" containerID="6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.678777 5035 scope.go:117] "RemoveContainer" containerID="a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.700479 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rkw2f"] Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.707588 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rkw2f"] Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.717054 5035 scope.go:117] "RemoveContainer" containerID="a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.781256 5035 scope.go:117] "RemoveContainer" containerID="6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c" Nov 24 19:39:19 crc kubenswrapper[5035]: E1124 19:39:19.781569 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c\": container with ID starting with 6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c not found: ID does not exist" containerID="6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.781597 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c"} err="failed to get container status \"6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c\": rpc error: code = NotFound desc = could not find container \"6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c\": container with ID starting with 6c8d023b5499da260e6af6b00f1229c51d36ae2b9f6ab7646321a168f7ed094c not found: ID does not exist" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.781617 5035 scope.go:117] "RemoveContainer" containerID="a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76" Nov 24 19:39:19 crc kubenswrapper[5035]: E1124 19:39:19.781810 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76\": container with ID starting with a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76 not found: ID does not exist" containerID="a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.781833 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76"} err="failed to get container status \"a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76\": rpc error: code = NotFound desc = could not find container \"a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76\": container with ID starting with a7f485f30430bf977d038354cb1040bd9951823723d194569477249ae99f9c76 not found: ID does not exist" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.781847 5035 scope.go:117] "RemoveContainer" containerID="a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2" Nov 24 19:39:19 crc kubenswrapper[5035]: E1124 19:39:19.782091 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2\": container with ID starting with a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2 not found: ID does not exist" containerID="a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2" Nov 24 19:39:19 crc kubenswrapper[5035]: I1124 19:39:19.782151 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2"} err="failed to get container status \"a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2\": rpc error: code = NotFound desc = could not find container \"a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2\": container with ID starting with a0eadc8868fd815c0cd064106cff32bbe06c79e89d3ccf94ca20e2ef177185f2 not found: ID does not exist" Nov 24 19:39:20 crc kubenswrapper[5035]: I1124 19:39:20.214961 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" path="/var/lib/kubelet/pods/fef91184-d485-4f47-abc8-7b7aac22d9d1/volumes" Nov 24 19:39:24 crc kubenswrapper[5035]: I1124 19:39:24.056111 5035 scope.go:117] "RemoveContainer" containerID="f9d14da07ccc04044087429f74e3a615ac05903701b026bdc52ac7caa9d187e3" Nov 24 19:39:45 crc kubenswrapper[5035]: I1124 19:39:45.234136 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:39:45 crc kubenswrapper[5035]: I1124 19:39:45.234864 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:40:15 crc kubenswrapper[5035]: I1124 19:40:15.234086 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:40:15 crc kubenswrapper[5035]: I1124 19:40:15.234591 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:40:15 crc kubenswrapper[5035]: I1124 19:40:15.234635 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:40:15 crc kubenswrapper[5035]: I1124 19:40:15.235273 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:40:15 crc kubenswrapper[5035]: I1124 19:40:15.235341 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" gracePeriod=600 Nov 24 19:40:15 crc kubenswrapper[5035]: E1124 19:40:15.366823 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:40:16 crc kubenswrapper[5035]: I1124 19:40:16.206266 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" exitCode=0 Nov 24 19:40:16 crc kubenswrapper[5035]: I1124 19:40:16.217652 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199"} Nov 24 19:40:16 crc kubenswrapper[5035]: I1124 19:40:16.217733 5035 scope.go:117] "RemoveContainer" containerID="5e56edadec21ff9db609cb1eb46281d662ed2c26ae0bd15afd911d69fb3661e1" Nov 24 19:40:16 crc kubenswrapper[5035]: I1124 19:40:16.218695 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:40:16 crc kubenswrapper[5035]: E1124 19:40:16.219207 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:40:30 crc kubenswrapper[5035]: I1124 19:40:30.200853 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:40:30 crc kubenswrapper[5035]: E1124 19:40:30.201828 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:40:41 crc kubenswrapper[5035]: I1124 19:40:41.199986 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:40:41 crc kubenswrapper[5035]: E1124 19:40:41.201031 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:40:54 crc kubenswrapper[5035]: I1124 19:40:54.200978 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:40:54 crc kubenswrapper[5035]: E1124 19:40:54.202168 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:41:08 crc kubenswrapper[5035]: I1124 19:41:08.201408 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:41:08 crc kubenswrapper[5035]: E1124 19:41:08.202680 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:41:19 crc kubenswrapper[5035]: I1124 19:41:19.200172 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:41:19 crc kubenswrapper[5035]: E1124 19:41:19.201033 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.251029 5035 scope.go:117] "RemoveContainer" containerID="059cc1ecdc7b970cadee29e2db971d102d536c8312737643c621bc251aae3f34" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.292389 5035 scope.go:117] "RemoveContainer" containerID="4856480726a423da01c3655a262332b2078a84b354378cce3495899f7c2960a4" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.324675 5035 scope.go:117] "RemoveContainer" containerID="79b3c0a93f919b8ed4792288c2a4d902a251d8e956acc16231886fbf335eef7f" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.343720 5035 scope.go:117] "RemoveContainer" containerID="0507d5c07ce9fb02a2b394735925f16b0843a9d3b866d91373902c03648c34e1" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.365462 5035 scope.go:117] "RemoveContainer" containerID="160a7bfaec1325b20eb24874f945f585a3af874dfd87562a5b8164bdd9d6d7e0" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.383069 5035 scope.go:117] "RemoveContainer" containerID="8fbb254c0db5b18f0d54fed5e8b6cf87978057bd5cd5afc94c18688ea509368a" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.412244 5035 scope.go:117] "RemoveContainer" containerID="4c0696c2b87e0bd41fbaed214a5bf7b1d0a902951f36ae9b00d9a44f93eb64a9" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.437180 5035 scope.go:117] "RemoveContainer" containerID="47ba28cd2e6da93f421d7f22340bdb55c53f765266a9e90869a0bf7643816343" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.456636 5035 scope.go:117] "RemoveContainer" containerID="7acc11f1e630c81f4a011df3c31e54e0128d9c6eea8c006ab4f842c1d719077f" Nov 24 19:41:24 crc kubenswrapper[5035]: I1124 19:41:24.482562 5035 scope.go:117] "RemoveContainer" containerID="fb6fa27cb03fbd7e03a0522a819abe8d9ca1d8d85dba3827283dceeb83a16910" Nov 24 19:41:29 crc kubenswrapper[5035]: I1124 19:41:29.037447 5035 generic.go:334] "Generic (PLEG): container finished" podID="5ee29809-634e-4c86-9605-5368bfc80b36" containerID="7107bfbed7b0fc114c81497ac87bcb2ddecd6b064f06f2d6a4d76b0601f54215" exitCode=0 Nov 24 19:41:29 crc kubenswrapper[5035]: I1124 19:41:29.037541 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" event={"ID":"5ee29809-634e-4c86-9605-5368bfc80b36","Type":"ContainerDied","Data":"7107bfbed7b0fc114c81497ac87bcb2ddecd6b064f06f2d6a4d76b0601f54215"} Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.205735 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:41:30 crc kubenswrapper[5035]: E1124 19:41:30.206454 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.524511 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.680487 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-inventory\") pod \"5ee29809-634e-4c86-9605-5368bfc80b36\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.680609 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-bootstrap-combined-ca-bundle\") pod \"5ee29809-634e-4c86-9605-5368bfc80b36\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.680880 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqm5q\" (UniqueName: \"kubernetes.io/projected/5ee29809-634e-4c86-9605-5368bfc80b36-kube-api-access-hqm5q\") pod \"5ee29809-634e-4c86-9605-5368bfc80b36\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.680951 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-ssh-key\") pod \"5ee29809-634e-4c86-9605-5368bfc80b36\" (UID: \"5ee29809-634e-4c86-9605-5368bfc80b36\") " Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.688244 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5ee29809-634e-4c86-9605-5368bfc80b36" (UID: "5ee29809-634e-4c86-9605-5368bfc80b36"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.688678 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee29809-634e-4c86-9605-5368bfc80b36-kube-api-access-hqm5q" (OuterVolumeSpecName: "kube-api-access-hqm5q") pod "5ee29809-634e-4c86-9605-5368bfc80b36" (UID: "5ee29809-634e-4c86-9605-5368bfc80b36"). InnerVolumeSpecName "kube-api-access-hqm5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.711752 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-inventory" (OuterVolumeSpecName: "inventory") pod "5ee29809-634e-4c86-9605-5368bfc80b36" (UID: "5ee29809-634e-4c86-9605-5368bfc80b36"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.713461 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5ee29809-634e-4c86-9605-5368bfc80b36" (UID: "5ee29809-634e-4c86-9605-5368bfc80b36"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.783480 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqm5q\" (UniqueName: \"kubernetes.io/projected/5ee29809-634e-4c86-9605-5368bfc80b36-kube-api-access-hqm5q\") on node \"crc\" DevicePath \"\"" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.783538 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.783562 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:41:30 crc kubenswrapper[5035]: I1124 19:41:30.783582 5035 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ee29809-634e-4c86-9605-5368bfc80b36-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.062919 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" event={"ID":"5ee29809-634e-4c86-9605-5368bfc80b36","Type":"ContainerDied","Data":"648e7fd4afe61661410ecc925718fb39f304a39a6c665bcf88bff1242605d566"} Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.063312 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="648e7fd4afe61661410ecc925718fb39f304a39a6c665bcf88bff1242605d566" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.063007 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.169451 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772"] Nov 24 19:41:31 crc kubenswrapper[5035]: E1124 19:41:31.170210 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="extract-utilities" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.170374 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="extract-utilities" Nov 24 19:41:31 crc kubenswrapper[5035]: E1124 19:41:31.171001 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="registry-server" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.171166 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="registry-server" Nov 24 19:41:31 crc kubenswrapper[5035]: E1124 19:41:31.171366 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee29809-634e-4c86-9605-5368bfc80b36" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.171548 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee29809-634e-4c86-9605-5368bfc80b36" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 19:41:31 crc kubenswrapper[5035]: E1124 19:41:31.171716 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="extract-content" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.171832 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="extract-content" Nov 24 19:41:31 crc kubenswrapper[5035]: E1124 19:41:31.171930 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="registry-server" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.172005 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="registry-server" Nov 24 19:41:31 crc kubenswrapper[5035]: E1124 19:41:31.172092 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="extract-utilities" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.172174 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="extract-utilities" Nov 24 19:41:31 crc kubenswrapper[5035]: E1124 19:41:31.172274 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="extract-content" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.172437 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="extract-content" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.172741 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef91184-d485-4f47-abc8-7b7aac22d9d1" containerName="registry-server" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.172862 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee29809-634e-4c86-9605-5368bfc80b36" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.172949 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="22efa5a8-e97f-4702-9a47-626b05cdb13a" containerName="registry-server" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.173810 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.176133 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.176470 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.177704 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.180832 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.217097 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772"] Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.305713 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbxgs\" (UniqueName: \"kubernetes.io/projected/bda65671-049d-4645-96fc-f6e582e70fcc-kube-api-access-bbxgs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.306115 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.306369 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.407667 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.407771 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbxgs\" (UniqueName: \"kubernetes.io/projected/bda65671-049d-4645-96fc-f6e582e70fcc-kube-api-access-bbxgs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.407838 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.414886 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.415199 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.432247 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbxgs\" (UniqueName: \"kubernetes.io/projected/bda65671-049d-4645-96fc-f6e582e70fcc-kube-api-access-bbxgs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9f772\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:31 crc kubenswrapper[5035]: I1124 19:41:31.497737 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:41:32 crc kubenswrapper[5035]: I1124 19:41:32.097862 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772"] Nov 24 19:41:32 crc kubenswrapper[5035]: I1124 19:41:32.106946 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 19:41:33 crc kubenswrapper[5035]: I1124 19:41:33.082662 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" event={"ID":"bda65671-049d-4645-96fc-f6e582e70fcc","Type":"ContainerStarted","Data":"400090af98e3ce470f10bb5d9bfcf877e2289c5688d7ca8eddfb4a3201f4bcaa"} Nov 24 19:41:33 crc kubenswrapper[5035]: I1124 19:41:33.083420 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" event={"ID":"bda65671-049d-4645-96fc-f6e582e70fcc","Type":"ContainerStarted","Data":"08b34f985c220aa07a6700572797b20e0bf387b49e6da40048379f4244920c37"} Nov 24 19:41:33 crc kubenswrapper[5035]: I1124 19:41:33.102750 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" podStartSLOduration=1.6656987239999999 podStartE2EDuration="2.102734761s" podCreationTimestamp="2025-11-24 19:41:31 +0000 UTC" firstStartedPulling="2025-11-24 19:41:32.10639366 +0000 UTC m=+1630.628899917" lastFinishedPulling="2025-11-24 19:41:32.543429697 +0000 UTC m=+1631.065935954" observedRunningTime="2025-11-24 19:41:33.094701493 +0000 UTC m=+1631.617207750" watchObservedRunningTime="2025-11-24 19:41:33.102734761 +0000 UTC m=+1631.625241018" Nov 24 19:41:43 crc kubenswrapper[5035]: I1124 19:41:43.200658 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:41:43 crc kubenswrapper[5035]: E1124 19:41:43.201744 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:41:52 crc kubenswrapper[5035]: I1124 19:41:52.059871 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pmbwf"] Nov 24 19:41:52 crc kubenswrapper[5035]: I1124 19:41:52.071023 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-edc5-account-create-bfvds"] Nov 24 19:41:52 crc kubenswrapper[5035]: I1124 19:41:52.080247 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pmbwf"] Nov 24 19:41:52 crc kubenswrapper[5035]: I1124 19:41:52.090393 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-edc5-account-create-bfvds"] Nov 24 19:41:52 crc kubenswrapper[5035]: I1124 19:41:52.211723 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a40bddb-4969-4100-a311-1fd2614bd9f8" path="/var/lib/kubelet/pods/1a40bddb-4969-4100-a311-1fd2614bd9f8/volumes" Nov 24 19:41:52 crc kubenswrapper[5035]: I1124 19:41:52.212281 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="395eef76-f4f1-4543-84b0-0d69abb77653" path="/var/lib/kubelet/pods/395eef76-f4f1-4543-84b0-0d69abb77653/volumes" Nov 24 19:41:56 crc kubenswrapper[5035]: I1124 19:41:56.045940 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-034b-account-create-g66v8"] Nov 24 19:41:56 crc kubenswrapper[5035]: I1124 19:41:56.062553 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-034b-account-create-g66v8"] Nov 24 19:41:56 crc kubenswrapper[5035]: I1124 19:41:56.217210 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="944131f2-13c8-4e25-b543-bc001f29a85a" path="/var/lib/kubelet/pods/944131f2-13c8-4e25-b543-bc001f29a85a/volumes" Nov 24 19:41:57 crc kubenswrapper[5035]: I1124 19:41:57.032779 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-z2rqz"] Nov 24 19:41:57 crc kubenswrapper[5035]: I1124 19:41:57.043849 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-vhfcj"] Nov 24 19:41:57 crc kubenswrapper[5035]: I1124 19:41:57.054922 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a61f-account-create-95bq4"] Nov 24 19:41:57 crc kubenswrapper[5035]: I1124 19:41:57.064425 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-z2rqz"] Nov 24 19:41:57 crc kubenswrapper[5035]: I1124 19:41:57.071617 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a61f-account-create-95bq4"] Nov 24 19:41:57 crc kubenswrapper[5035]: I1124 19:41:57.079696 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-vhfcj"] Nov 24 19:41:58 crc kubenswrapper[5035]: I1124 19:41:58.200406 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:41:58 crc kubenswrapper[5035]: E1124 19:41:58.200862 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:41:58 crc kubenswrapper[5035]: I1124 19:41:58.215797 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="817e8206-e355-4f96-b495-a2afab24f1c4" path="/var/lib/kubelet/pods/817e8206-e355-4f96-b495-a2afab24f1c4/volumes" Nov 24 19:41:58 crc kubenswrapper[5035]: I1124 19:41:58.217096 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c373792a-2d1a-4030-bbb4-3fa4af555411" path="/var/lib/kubelet/pods/c373792a-2d1a-4030-bbb4-3fa4af555411/volumes" Nov 24 19:41:58 crc kubenswrapper[5035]: I1124 19:41:58.218619 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f3e40f-fd65-4c1c-b3e9-7d2158a97038" path="/var/lib/kubelet/pods/e8f3e40f-fd65-4c1c-b3e9-7d2158a97038/volumes" Nov 24 19:42:10 crc kubenswrapper[5035]: I1124 19:42:10.199936 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:42:10 crc kubenswrapper[5035]: E1124 19:42:10.200934 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.045044 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vrrsl"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.070909 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dgz8h"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.079976 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c419-account-create-nxsz2"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.089745 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dgz8h"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.096844 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vrrsl"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.103320 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c419-account-create-nxsz2"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.110481 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d325-account-create-dwc79"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.116427 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d325-account-create-dwc79"] Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.216751 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fcf962-da19-4161-8eee-aecde6615ba4" path="/var/lib/kubelet/pods/68fcf962-da19-4161-8eee-aecde6615ba4/volumes" Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.217982 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80412713-2ef8-4499-ae54-2c6aebac5377" path="/var/lib/kubelet/pods/80412713-2ef8-4499-ae54-2c6aebac5377/volumes" Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.219203 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d84b9e7-688c-4926-bdce-1581f89d4875" path="/var/lib/kubelet/pods/8d84b9e7-688c-4926-bdce-1581f89d4875/volumes" Nov 24 19:42:14 crc kubenswrapper[5035]: I1124 19:42:14.220385 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e080c702-5f10-4d8a-a192-07466d0d7644" path="/var/lib/kubelet/pods/e080c702-5f10-4d8a-a192-07466d0d7644/volumes" Nov 24 19:42:20 crc kubenswrapper[5035]: I1124 19:42:20.041622 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f85b-account-create-zpj9j"] Nov 24 19:42:20 crc kubenswrapper[5035]: I1124 19:42:20.053266 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f85b-account-create-zpj9j"] Nov 24 19:42:20 crc kubenswrapper[5035]: I1124 19:42:20.064465 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-cxbtl"] Nov 24 19:42:20 crc kubenswrapper[5035]: I1124 19:42:20.076888 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-cxbtl"] Nov 24 19:42:20 crc kubenswrapper[5035]: I1124 19:42:20.213189 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeff8377-e685-4e82-ad78-b9b99f9e1bc0" path="/var/lib/kubelet/pods/aeff8377-e685-4e82-ad78-b9b99f9e1bc0/volumes" Nov 24 19:42:20 crc kubenswrapper[5035]: I1124 19:42:20.213954 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da1e9bbe-02f8-4eda-be3d-bf882329f44c" path="/var/lib/kubelet/pods/da1e9bbe-02f8-4eda-be3d-bf882329f44c/volumes" Nov 24 19:42:23 crc kubenswrapper[5035]: I1124 19:42:23.030849 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-qrcbh"] Nov 24 19:42:23 crc kubenswrapper[5035]: I1124 19:42:23.047147 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-qrcbh"] Nov 24 19:42:23 crc kubenswrapper[5035]: I1124 19:42:23.200609 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:42:23 crc kubenswrapper[5035]: E1124 19:42:23.200945 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.219353 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="030eac16-e8f0-4a3c-9d1e-588cc25d6ebe" path="/var/lib/kubelet/pods/030eac16-e8f0-4a3c-9d1e-588cc25d6ebe/volumes" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.594995 5035 scope.go:117] "RemoveContainer" containerID="0cb2d0ed629ff5752b787fb0ddb242d5f52d83836f94f0c0ed401ad9638ba9aa" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.634846 5035 scope.go:117] "RemoveContainer" containerID="805534dd8880ebe48abf9b3a13b871ab38cd236f5d2fb19d4e1a9388254c77d9" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.685848 5035 scope.go:117] "RemoveContainer" containerID="8521c323661c784a4292e298b4f92f72438e6d82650f8d6785f444a848ff7881" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.725170 5035 scope.go:117] "RemoveContainer" containerID="a139d5c50022bea46fcaabbf66a045cfd73e2839694412004c8747dfe942cf21" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.760529 5035 scope.go:117] "RemoveContainer" containerID="d8da86f1d1cd555086611c05df2b0b0c4b17b03065becb8d4fcd1bf9d2b66b32" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.797743 5035 scope.go:117] "RemoveContainer" containerID="5a6251900743edf35f1d5fcf6eba33637668b21e2d087572af49e2366a4d91a3" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.839388 5035 scope.go:117] "RemoveContainer" containerID="cd14a78ee5631624f60504e02515fdfe74264d2f2eb42000a2153830c2f52373" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.861252 5035 scope.go:117] "RemoveContainer" containerID="2b65dff80626f7d183a900f4cf6dc685afc94965cc5638d12d4419aa23d9d680" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.879685 5035 scope.go:117] "RemoveContainer" containerID="492f865652f7fcd134fb8860f0d4568381a96e467551d261b344b91454c4f9f9" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.897149 5035 scope.go:117] "RemoveContainer" containerID="4cce725c77ece9593b000a253ae81b3eb7df00c2a24ec012d0b1f9d34272c134" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.940396 5035 scope.go:117] "RemoveContainer" containerID="2ca0cee62ad3f7a297bfcda9bc92aef39dfd13f7ec4bc0c6dbe1f9d0d6ca24be" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.964895 5035 scope.go:117] "RemoveContainer" containerID="0137522e2f0147317312b1acc52d8f5f9815e67b3584dbb91490d969b14317fb" Nov 24 19:42:24 crc kubenswrapper[5035]: I1124 19:42:24.985838 5035 scope.go:117] "RemoveContainer" containerID="f54a44ccc3e2b2712001515b65b9420624b3c71fbf91f2786740003a91fd887e" Nov 24 19:42:26 crc kubenswrapper[5035]: I1124 19:42:26.063238 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-9jwdn"] Nov 24 19:42:26 crc kubenswrapper[5035]: I1124 19:42:26.079695 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-9jwdn"] Nov 24 19:42:26 crc kubenswrapper[5035]: I1124 19:42:26.209435 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ba53ebf-bae3-4629-af0b-8637e6491515" path="/var/lib/kubelet/pods/7ba53ebf-bae3-4629-af0b-8637e6491515/volumes" Nov 24 19:42:37 crc kubenswrapper[5035]: I1124 19:42:37.202669 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:42:37 crc kubenswrapper[5035]: E1124 19:42:37.203715 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:42:46 crc kubenswrapper[5035]: I1124 19:42:46.739105 5035 generic.go:334] "Generic (PLEG): container finished" podID="bda65671-049d-4645-96fc-f6e582e70fcc" containerID="400090af98e3ce470f10bb5d9bfcf877e2289c5688d7ca8eddfb4a3201f4bcaa" exitCode=0 Nov 24 19:42:46 crc kubenswrapper[5035]: I1124 19:42:46.739451 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" event={"ID":"bda65671-049d-4645-96fc-f6e582e70fcc","Type":"ContainerDied","Data":"400090af98e3ce470f10bb5d9bfcf877e2289c5688d7ca8eddfb4a3201f4bcaa"} Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.124551 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.243380 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbxgs\" (UniqueName: \"kubernetes.io/projected/bda65671-049d-4645-96fc-f6e582e70fcc-kube-api-access-bbxgs\") pod \"bda65671-049d-4645-96fc-f6e582e70fcc\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.243442 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-ssh-key\") pod \"bda65671-049d-4645-96fc-f6e582e70fcc\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.243464 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-inventory\") pod \"bda65671-049d-4645-96fc-f6e582e70fcc\" (UID: \"bda65671-049d-4645-96fc-f6e582e70fcc\") " Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.249296 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bda65671-049d-4645-96fc-f6e582e70fcc-kube-api-access-bbxgs" (OuterVolumeSpecName: "kube-api-access-bbxgs") pod "bda65671-049d-4645-96fc-f6e582e70fcc" (UID: "bda65671-049d-4645-96fc-f6e582e70fcc"). InnerVolumeSpecName "kube-api-access-bbxgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.269614 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bda65671-049d-4645-96fc-f6e582e70fcc" (UID: "bda65671-049d-4645-96fc-f6e582e70fcc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.287051 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-inventory" (OuterVolumeSpecName: "inventory") pod "bda65671-049d-4645-96fc-f6e582e70fcc" (UID: "bda65671-049d-4645-96fc-f6e582e70fcc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.345729 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbxgs\" (UniqueName: \"kubernetes.io/projected/bda65671-049d-4645-96fc-f6e582e70fcc-kube-api-access-bbxgs\") on node \"crc\" DevicePath \"\"" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.345774 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.345788 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bda65671-049d-4645-96fc-f6e582e70fcc-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.754881 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" event={"ID":"bda65671-049d-4645-96fc-f6e582e70fcc","Type":"ContainerDied","Data":"08b34f985c220aa07a6700572797b20e0bf387b49e6da40048379f4244920c37"} Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.755166 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08b34f985c220aa07a6700572797b20e0bf387b49e6da40048379f4244920c37" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.754934 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.881787 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94"] Nov 24 19:42:48 crc kubenswrapper[5035]: E1124 19:42:48.882512 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda65671-049d-4645-96fc-f6e582e70fcc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.882618 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda65671-049d-4645-96fc-f6e582e70fcc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.882955 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bda65671-049d-4645-96fc-f6e582e70fcc" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.883907 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.886111 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.886276 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.886686 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.886831 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.893024 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94"] Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.956142 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcktd\" (UniqueName: \"kubernetes.io/projected/878700e6-143e-4bb4-a300-13fd3ce143ac-kube-api-access-kcktd\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.956211 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:48 crc kubenswrapper[5035]: I1124 19:42:48.956284 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:49 crc kubenswrapper[5035]: I1124 19:42:49.058088 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:49 crc kubenswrapper[5035]: I1124 19:42:49.058258 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcktd\" (UniqueName: \"kubernetes.io/projected/878700e6-143e-4bb4-a300-13fd3ce143ac-kube-api-access-kcktd\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:49 crc kubenswrapper[5035]: I1124 19:42:49.058339 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:49 crc kubenswrapper[5035]: I1124 19:42:49.062723 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:49 crc kubenswrapper[5035]: I1124 19:42:49.071002 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:49 crc kubenswrapper[5035]: I1124 19:42:49.084161 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcktd\" (UniqueName: \"kubernetes.io/projected/878700e6-143e-4bb4-a300-13fd3ce143ac-kube-api-access-kcktd\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-26d94\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:49 crc kubenswrapper[5035]: I1124 19:42:49.202544 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:50 crc kubenswrapper[5035]: I1124 19:42:49.761227 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94"] Nov 24 19:42:50 crc kubenswrapper[5035]: I1124 19:42:50.780763 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" event={"ID":"878700e6-143e-4bb4-a300-13fd3ce143ac","Type":"ContainerStarted","Data":"2d964ff0dc045abf4c4834bf3939982aea08ab94133bed49ec875a2f9ea9e82a"} Nov 24 19:42:50 crc kubenswrapper[5035]: I1124 19:42:50.781499 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" event={"ID":"878700e6-143e-4bb4-a300-13fd3ce143ac","Type":"ContainerStarted","Data":"c9a6473623c7f9ea636b8f176c254ae703a02ae830e9428879db996eec85b6a2"} Nov 24 19:42:50 crc kubenswrapper[5035]: I1124 19:42:50.798115 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" podStartSLOduration=2.357284084 podStartE2EDuration="2.798090863s" podCreationTimestamp="2025-11-24 19:42:48 +0000 UTC" firstStartedPulling="2025-11-24 19:42:49.775055049 +0000 UTC m=+1708.297561306" lastFinishedPulling="2025-11-24 19:42:50.215861818 +0000 UTC m=+1708.738368085" observedRunningTime="2025-11-24 19:42:50.795908064 +0000 UTC m=+1709.318414341" watchObservedRunningTime="2025-11-24 19:42:50.798090863 +0000 UTC m=+1709.320597140" Nov 24 19:42:51 crc kubenswrapper[5035]: I1124 19:42:51.200368 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:42:51 crc kubenswrapper[5035]: E1124 19:42:51.200670 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:42:55 crc kubenswrapper[5035]: I1124 19:42:55.835763 5035 generic.go:334] "Generic (PLEG): container finished" podID="878700e6-143e-4bb4-a300-13fd3ce143ac" containerID="2d964ff0dc045abf4c4834bf3939982aea08ab94133bed49ec875a2f9ea9e82a" exitCode=0 Nov 24 19:42:55 crc kubenswrapper[5035]: I1124 19:42:55.835868 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" event={"ID":"878700e6-143e-4bb4-a300-13fd3ce143ac","Type":"ContainerDied","Data":"2d964ff0dc045abf4c4834bf3939982aea08ab94133bed49ec875a2f9ea9e82a"} Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.259835 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.308055 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcktd\" (UniqueName: \"kubernetes.io/projected/878700e6-143e-4bb4-a300-13fd3ce143ac-kube-api-access-kcktd\") pod \"878700e6-143e-4bb4-a300-13fd3ce143ac\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.308230 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory\") pod \"878700e6-143e-4bb4-a300-13fd3ce143ac\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.308380 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-ssh-key\") pod \"878700e6-143e-4bb4-a300-13fd3ce143ac\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.313764 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/878700e6-143e-4bb4-a300-13fd3ce143ac-kube-api-access-kcktd" (OuterVolumeSpecName: "kube-api-access-kcktd") pod "878700e6-143e-4bb4-a300-13fd3ce143ac" (UID: "878700e6-143e-4bb4-a300-13fd3ce143ac"). InnerVolumeSpecName "kube-api-access-kcktd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:42:57 crc kubenswrapper[5035]: E1124 19:42:57.338624 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory podName:878700e6-143e-4bb4-a300-13fd3ce143ac nodeName:}" failed. No retries permitted until 2025-11-24 19:42:57.838589631 +0000 UTC m=+1716.361095898 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory") pod "878700e6-143e-4bb4-a300-13fd3ce143ac" (UID: "878700e6-143e-4bb4-a300-13fd3ce143ac") : error deleting /var/lib/kubelet/pods/878700e6-143e-4bb4-a300-13fd3ce143ac/volume-subpaths: remove /var/lib/kubelet/pods/878700e6-143e-4bb4-a300-13fd3ce143ac/volume-subpaths: no such file or directory Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.341707 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "878700e6-143e-4bb4-a300-13fd3ce143ac" (UID: "878700e6-143e-4bb4-a300-13fd3ce143ac"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.410127 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcktd\" (UniqueName: \"kubernetes.io/projected/878700e6-143e-4bb4-a300-13fd3ce143ac-kube-api-access-kcktd\") on node \"crc\" DevicePath \"\"" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.410165 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.862440 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" event={"ID":"878700e6-143e-4bb4-a300-13fd3ce143ac","Type":"ContainerDied","Data":"c9a6473623c7f9ea636b8f176c254ae703a02ae830e9428879db996eec85b6a2"} Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.862493 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9a6473623c7f9ea636b8f176c254ae703a02ae830e9428879db996eec85b6a2" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.862874 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.920412 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory\") pod \"878700e6-143e-4bb4-a300-13fd3ce143ac\" (UID: \"878700e6-143e-4bb4-a300-13fd3ce143ac\") " Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.925534 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory" (OuterVolumeSpecName: "inventory") pod "878700e6-143e-4bb4-a300-13fd3ce143ac" (UID: "878700e6-143e-4bb4-a300-13fd3ce143ac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.931360 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn"] Nov 24 19:42:57 crc kubenswrapper[5035]: E1124 19:42:57.931715 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="878700e6-143e-4bb4-a300-13fd3ce143ac" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.931727 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="878700e6-143e-4bb4-a300-13fd3ce143ac" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.931913 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="878700e6-143e-4bb4-a300-13fd3ce143ac" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.932553 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:57 crc kubenswrapper[5035]: I1124 19:42:57.947271 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn"] Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.022345 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.022605 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.022792 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjb4h\" (UniqueName: \"kubernetes.io/projected/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-kube-api-access-tjb4h\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.022868 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/878700e6-143e-4bb4-a300-13fd3ce143ac-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.124546 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjb4h\" (UniqueName: \"kubernetes.io/projected/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-kube-api-access-tjb4h\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.124654 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.124734 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.128886 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.130245 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.140194 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjb4h\" (UniqueName: \"kubernetes.io/projected/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-kube-api-access-tjb4h\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-d8dhn\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.307316 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:42:58 crc kubenswrapper[5035]: I1124 19:42:58.907111 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn"] Nov 24 19:42:58 crc kubenswrapper[5035]: W1124 19:42:58.912180 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec6e4be0_9f98_41d1_8bba_563fe9b39c78.slice/crio-3e9dfad163a195ed9bd27caff461356dd5aad4410c139e027dab05bd7495fb26 WatchSource:0}: Error finding container 3e9dfad163a195ed9bd27caff461356dd5aad4410c139e027dab05bd7495fb26: Status 404 returned error can't find the container with id 3e9dfad163a195ed9bd27caff461356dd5aad4410c139e027dab05bd7495fb26 Nov 24 19:42:59 crc kubenswrapper[5035]: I1124 19:42:59.888662 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" event={"ID":"ec6e4be0-9f98-41d1-8bba-563fe9b39c78","Type":"ContainerStarted","Data":"3d451ec7a73ffb3c6ebc60f845290c24e9b8c584d9d3251c9733ea3586004d03"} Nov 24 19:42:59 crc kubenswrapper[5035]: I1124 19:42:59.889233 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" event={"ID":"ec6e4be0-9f98-41d1-8bba-563fe9b39c78","Type":"ContainerStarted","Data":"3e9dfad163a195ed9bd27caff461356dd5aad4410c139e027dab05bd7495fb26"} Nov 24 19:42:59 crc kubenswrapper[5035]: I1124 19:42:59.910598 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" podStartSLOduration=2.482599308 podStartE2EDuration="2.910564007s" podCreationTimestamp="2025-11-24 19:42:57 +0000 UTC" firstStartedPulling="2025-11-24 19:42:58.914791523 +0000 UTC m=+1717.437297800" lastFinishedPulling="2025-11-24 19:42:59.342756242 +0000 UTC m=+1717.865262499" observedRunningTime="2025-11-24 19:42:59.902647802 +0000 UTC m=+1718.425154059" watchObservedRunningTime="2025-11-24 19:42:59.910564007 +0000 UTC m=+1718.433070304" Nov 24 19:43:03 crc kubenswrapper[5035]: I1124 19:43:03.200654 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:43:03 crc kubenswrapper[5035]: E1124 19:43:03.201650 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:43:08 crc kubenswrapper[5035]: I1124 19:43:08.045151 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kfcrg"] Nov 24 19:43:08 crc kubenswrapper[5035]: I1124 19:43:08.058728 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kfcrg"] Nov 24 19:43:08 crc kubenswrapper[5035]: I1124 19:43:08.213167 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="391f762a-cd72-4011-abc8-6efc3fce4661" path="/var/lib/kubelet/pods/391f762a-cd72-4011-abc8-6efc3fce4661/volumes" Nov 24 19:43:09 crc kubenswrapper[5035]: I1124 19:43:09.036837 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-h76mn"] Nov 24 19:43:09 crc kubenswrapper[5035]: I1124 19:43:09.043554 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-h76mn"] Nov 24 19:43:10 crc kubenswrapper[5035]: I1124 19:43:10.030207 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xb7gt"] Nov 24 19:43:10 crc kubenswrapper[5035]: I1124 19:43:10.037317 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xb7gt"] Nov 24 19:43:10 crc kubenswrapper[5035]: I1124 19:43:10.213741 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="205bbe21-12fb-4cdf-bf2d-09008085f221" path="/var/lib/kubelet/pods/205bbe21-12fb-4cdf-bf2d-09008085f221/volumes" Nov 24 19:43:10 crc kubenswrapper[5035]: I1124 19:43:10.215517 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="764ce995-4314-4c29-8b00-dc264344edf0" path="/var/lib/kubelet/pods/764ce995-4314-4c29-8b00-dc264344edf0/volumes" Nov 24 19:43:15 crc kubenswrapper[5035]: I1124 19:43:15.201052 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:43:15 crc kubenswrapper[5035]: E1124 19:43:15.202388 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:43:19 crc kubenswrapper[5035]: I1124 19:43:19.046083 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-nbbjc"] Nov 24 19:43:19 crc kubenswrapper[5035]: I1124 19:43:19.055540 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-nbbjc"] Nov 24 19:43:20 crc kubenswrapper[5035]: I1124 19:43:20.214153 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9794ab-e4e3-4af0-a689-6ef761d1fe4f" path="/var/lib/kubelet/pods/0c9794ab-e4e3-4af0-a689-6ef761d1fe4f/volumes" Nov 24 19:43:22 crc kubenswrapper[5035]: I1124 19:43:22.029282 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-g7nvj"] Nov 24 19:43:22 crc kubenswrapper[5035]: I1124 19:43:22.046151 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-g7nvj"] Nov 24 19:43:22 crc kubenswrapper[5035]: I1124 19:43:22.224809 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c11a04cb-bd10-49c8-affa-5189af979040" path="/var/lib/kubelet/pods/c11a04cb-bd10-49c8-affa-5189af979040/volumes" Nov 24 19:43:25 crc kubenswrapper[5035]: I1124 19:43:25.175887 5035 scope.go:117] "RemoveContainer" containerID="2ee8fc07fd677e9f8fc0399a5dab162e4cd2870a2b6cf1bf35ff520a6563ad3f" Nov 24 19:43:25 crc kubenswrapper[5035]: I1124 19:43:25.237337 5035 scope.go:117] "RemoveContainer" containerID="1ae6fe50bc26642ebd076d87c538a9201743427dfe2b961d2e5836739ba4d120" Nov 24 19:43:25 crc kubenswrapper[5035]: I1124 19:43:25.283054 5035 scope.go:117] "RemoveContainer" containerID="21dabb6c2c2e93d18439131b14ef79c29831cb76666b75d28f8ff63a3ce15098" Nov 24 19:43:25 crc kubenswrapper[5035]: I1124 19:43:25.348664 5035 scope.go:117] "RemoveContainer" containerID="700413af8825bd3f07de635668905592f6f45b1f467cfbebf94e1ef465dfacd9" Nov 24 19:43:25 crc kubenswrapper[5035]: I1124 19:43:25.387137 5035 scope.go:117] "RemoveContainer" containerID="4c6bb91f085a9a7f1b6c300c57b5ba3691a074db9a73773a414462dcd34bb968" Nov 24 19:43:25 crc kubenswrapper[5035]: I1124 19:43:25.435323 5035 scope.go:117] "RemoveContainer" containerID="37ef83088d4ee74d7be789192a3e95549cb61c54efe81b083affc749f3aff13c" Nov 24 19:43:30 crc kubenswrapper[5035]: I1124 19:43:30.200488 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:43:30 crc kubenswrapper[5035]: E1124 19:43:30.201922 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:43:43 crc kubenswrapper[5035]: I1124 19:43:43.292362 5035 generic.go:334] "Generic (PLEG): container finished" podID="ec6e4be0-9f98-41d1-8bba-563fe9b39c78" containerID="3d451ec7a73ffb3c6ebc60f845290c24e9b8c584d9d3251c9733ea3586004d03" exitCode=0 Nov 24 19:43:43 crc kubenswrapper[5035]: I1124 19:43:43.292479 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" event={"ID":"ec6e4be0-9f98-41d1-8bba-563fe9b39c78","Type":"ContainerDied","Data":"3d451ec7a73ffb3c6ebc60f845290c24e9b8c584d9d3251c9733ea3586004d03"} Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.200966 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:43:44 crc kubenswrapper[5035]: E1124 19:43:44.201489 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.733643 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.930404 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-inventory\") pod \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.930532 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjb4h\" (UniqueName: \"kubernetes.io/projected/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-kube-api-access-tjb4h\") pod \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.930716 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-ssh-key\") pod \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\" (UID: \"ec6e4be0-9f98-41d1-8bba-563fe9b39c78\") " Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.937673 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-kube-api-access-tjb4h" (OuterVolumeSpecName: "kube-api-access-tjb4h") pod "ec6e4be0-9f98-41d1-8bba-563fe9b39c78" (UID: "ec6e4be0-9f98-41d1-8bba-563fe9b39c78"). InnerVolumeSpecName "kube-api-access-tjb4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.959583 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec6e4be0-9f98-41d1-8bba-563fe9b39c78" (UID: "ec6e4be0-9f98-41d1-8bba-563fe9b39c78"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:43:44 crc kubenswrapper[5035]: I1124 19:43:44.962023 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-inventory" (OuterVolumeSpecName: "inventory") pod "ec6e4be0-9f98-41d1-8bba-563fe9b39c78" (UID: "ec6e4be0-9f98-41d1-8bba-563fe9b39c78"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.033988 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.034267 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjb4h\" (UniqueName: \"kubernetes.io/projected/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-kube-api-access-tjb4h\") on node \"crc\" DevicePath \"\"" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.034376 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6e4be0-9f98-41d1-8bba-563fe9b39c78-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.316166 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" event={"ID":"ec6e4be0-9f98-41d1-8bba-563fe9b39c78","Type":"ContainerDied","Data":"3e9dfad163a195ed9bd27caff461356dd5aad4410c139e027dab05bd7495fb26"} Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.316213 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e9dfad163a195ed9bd27caff461356dd5aad4410c139e027dab05bd7495fb26" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.316281 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.416819 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8"] Nov 24 19:43:45 crc kubenswrapper[5035]: E1124 19:43:45.417207 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6e4be0-9f98-41d1-8bba-563fe9b39c78" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.417225 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6e4be0-9f98-41d1-8bba-563fe9b39c78" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.417404 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec6e4be0-9f98-41d1-8bba-563fe9b39c78" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.418085 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.420505 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.421460 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.421671 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.421764 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.435157 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8"] Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.544856 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghl8d\" (UniqueName: \"kubernetes.io/projected/eb570109-faff-4cd5-97ac-b9d0ac5398fb-kube-api-access-ghl8d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.545005 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.545108 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.647215 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghl8d\" (UniqueName: \"kubernetes.io/projected/eb570109-faff-4cd5-97ac-b9d0ac5398fb-kube-api-access-ghl8d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.647661 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.648211 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.651826 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.651879 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.667550 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghl8d\" (UniqueName: \"kubernetes.io/projected/eb570109-faff-4cd5-97ac-b9d0ac5398fb-kube-api-access-ghl8d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:45 crc kubenswrapper[5035]: I1124 19:43:45.738878 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:46 crc kubenswrapper[5035]: I1124 19:43:46.256699 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8"] Nov 24 19:43:46 crc kubenswrapper[5035]: I1124 19:43:46.325979 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" event={"ID":"eb570109-faff-4cd5-97ac-b9d0ac5398fb","Type":"ContainerStarted","Data":"8646173a7870e546a998323ad4a6faf0dca5f1dde01b62aedcef5d0d4d9394b0"} Nov 24 19:43:47 crc kubenswrapper[5035]: I1124 19:43:47.336121 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" event={"ID":"eb570109-faff-4cd5-97ac-b9d0ac5398fb","Type":"ContainerStarted","Data":"7a6b96b6aff7d81089fb7bc81d6f7699e44fd904b924bddda69e1d51a6f701b5"} Nov 24 19:43:47 crc kubenswrapper[5035]: I1124 19:43:47.363085 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" podStartSLOduration=1.79803107 podStartE2EDuration="2.3630652s" podCreationTimestamp="2025-11-24 19:43:45 +0000 UTC" firstStartedPulling="2025-11-24 19:43:46.259667706 +0000 UTC m=+1764.782173973" lastFinishedPulling="2025-11-24 19:43:46.824701846 +0000 UTC m=+1765.347208103" observedRunningTime="2025-11-24 19:43:47.354115806 +0000 UTC m=+1765.876622063" watchObservedRunningTime="2025-11-24 19:43:47.3630652 +0000 UTC m=+1765.885571457" Nov 24 19:43:51 crc kubenswrapper[5035]: I1124 19:43:51.369825 5035 generic.go:334] "Generic (PLEG): container finished" podID="eb570109-faff-4cd5-97ac-b9d0ac5398fb" containerID="7a6b96b6aff7d81089fb7bc81d6f7699e44fd904b924bddda69e1d51a6f701b5" exitCode=0 Nov 24 19:43:51 crc kubenswrapper[5035]: I1124 19:43:51.369982 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" event={"ID":"eb570109-faff-4cd5-97ac-b9d0ac5398fb","Type":"ContainerDied","Data":"7a6b96b6aff7d81089fb7bc81d6f7699e44fd904b924bddda69e1d51a6f701b5"} Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.780960 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.892130 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-ssh-key\") pod \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.892907 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-inventory\") pod \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.893023 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghl8d\" (UniqueName: \"kubernetes.io/projected/eb570109-faff-4cd5-97ac-b9d0ac5398fb-kube-api-access-ghl8d\") pod \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\" (UID: \"eb570109-faff-4cd5-97ac-b9d0ac5398fb\") " Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.910457 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb570109-faff-4cd5-97ac-b9d0ac5398fb-kube-api-access-ghl8d" (OuterVolumeSpecName: "kube-api-access-ghl8d") pod "eb570109-faff-4cd5-97ac-b9d0ac5398fb" (UID: "eb570109-faff-4cd5-97ac-b9d0ac5398fb"). InnerVolumeSpecName "kube-api-access-ghl8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.927960 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-inventory" (OuterVolumeSpecName: "inventory") pod "eb570109-faff-4cd5-97ac-b9d0ac5398fb" (UID: "eb570109-faff-4cd5-97ac-b9d0ac5398fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.938088 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eb570109-faff-4cd5-97ac-b9d0ac5398fb" (UID: "eb570109-faff-4cd5-97ac-b9d0ac5398fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.997044 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.997076 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb570109-faff-4cd5-97ac-b9d0ac5398fb-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:43:52 crc kubenswrapper[5035]: I1124 19:43:52.997107 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghl8d\" (UniqueName: \"kubernetes.io/projected/eb570109-faff-4cd5-97ac-b9d0ac5398fb-kube-api-access-ghl8d\") on node \"crc\" DevicePath \"\"" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.397461 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" event={"ID":"eb570109-faff-4cd5-97ac-b9d0ac5398fb","Type":"ContainerDied","Data":"8646173a7870e546a998323ad4a6faf0dca5f1dde01b62aedcef5d0d4d9394b0"} Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.397497 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8646173a7870e546a998323ad4a6faf0dca5f1dde01b62aedcef5d0d4d9394b0" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.397579 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.470841 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx"] Nov 24 19:43:53 crc kubenswrapper[5035]: E1124 19:43:53.471561 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb570109-faff-4cd5-97ac-b9d0ac5398fb" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.471651 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb570109-faff-4cd5-97ac-b9d0ac5398fb" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.472019 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb570109-faff-4cd5-97ac-b9d0ac5398fb" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.472874 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.478574 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx"] Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.478700 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.478918 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.479087 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.479096 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.505879 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.506177 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.506277 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj4nd\" (UniqueName: \"kubernetes.io/projected/945f0c1b-86e7-4759-96fb-cc5278b03215-kube-api-access-gj4nd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.608115 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.608217 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.608250 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj4nd\" (UniqueName: \"kubernetes.io/projected/945f0c1b-86e7-4759-96fb-cc5278b03215-kube-api-access-gj4nd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.612019 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.612123 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.626658 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj4nd\" (UniqueName: \"kubernetes.io/projected/945f0c1b-86e7-4759-96fb-cc5278b03215-kube-api-access-gj4nd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:53 crc kubenswrapper[5035]: I1124 19:43:53.791905 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:43:54 crc kubenswrapper[5035]: I1124 19:43:54.353528 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx"] Nov 24 19:43:54 crc kubenswrapper[5035]: I1124 19:43:54.405417 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" event={"ID":"945f0c1b-86e7-4759-96fb-cc5278b03215","Type":"ContainerStarted","Data":"add4b72a8fe607e58359a899a2397f87d11017c94b59c13f13c941750a9d3e4e"} Nov 24 19:43:55 crc kubenswrapper[5035]: I1124 19:43:55.417743 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" event={"ID":"945f0c1b-86e7-4759-96fb-cc5278b03215","Type":"ContainerStarted","Data":"1ae38fe2079d3cb2feb21ffbbe6deb6997d914223fdc20ece9b53910be21b0c3"} Nov 24 19:43:55 crc kubenswrapper[5035]: I1124 19:43:55.437815 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" podStartSLOduration=2.006532637 podStartE2EDuration="2.4377633s" podCreationTimestamp="2025-11-24 19:43:53 +0000 UTC" firstStartedPulling="2025-11-24 19:43:54.364982931 +0000 UTC m=+1772.887489188" lastFinishedPulling="2025-11-24 19:43:54.796213594 +0000 UTC m=+1773.318719851" observedRunningTime="2025-11-24 19:43:55.434765038 +0000 UTC m=+1773.957271295" watchObservedRunningTime="2025-11-24 19:43:55.4377633 +0000 UTC m=+1773.960269557" Nov 24 19:43:59 crc kubenswrapper[5035]: I1124 19:43:59.199946 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:43:59 crc kubenswrapper[5035]: E1124 19:43:59.200680 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.044357 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-6317-account-create-j892b"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.051898 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-7s5zw"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.058847 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6f8f-account-create-57tjg"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.068450 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-6dq77"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.074915 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-w4w56"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.081124 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-6e5c-account-create-vkrb7"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.087260 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6f8f-account-create-57tjg"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.093427 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-6317-account-create-j892b"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.099364 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-6e5c-account-create-vkrb7"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.105568 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-7s5zw"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.111642 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-6dq77"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.117665 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-w4w56"] Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.215331 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0760f5c2-67dd-43ee-918c-75156c940da8" path="/var/lib/kubelet/pods/0760f5c2-67dd-43ee-918c-75156c940da8/volumes" Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.231222 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e427ed7-2510-4cf8-b9f2-bb30afa79eaa" path="/var/lib/kubelet/pods/0e427ed7-2510-4cf8-b9f2-bb30afa79eaa/volumes" Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.234108 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11977f76-cda4-437d-8ced-e061a8471b5b" path="/var/lib/kubelet/pods/11977f76-cda4-437d-8ced-e061a8471b5b/volumes" Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.235620 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ec91d4e-a697-46e8-b00e-b7dab491189d" path="/var/lib/kubelet/pods/3ec91d4e-a697-46e8-b00e-b7dab491189d/volumes" Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.237545 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b75a4d05-f082-474e-b25f-f5889ed418b4" path="/var/lib/kubelet/pods/b75a4d05-f082-474e-b25f-f5889ed418b4/volumes" Nov 24 19:44:00 crc kubenswrapper[5035]: I1124 19:44:00.238568 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d88bc46f-ae00-4877-b99c-478f8891e14e" path="/var/lib/kubelet/pods/d88bc46f-ae00-4877-b99c-478f8891e14e/volumes" Nov 24 19:44:12 crc kubenswrapper[5035]: I1124 19:44:12.208242 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:44:12 crc kubenswrapper[5035]: E1124 19:44:12.209075 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:44:24 crc kubenswrapper[5035]: I1124 19:44:24.200901 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:44:24 crc kubenswrapper[5035]: E1124 19:44:24.201598 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:44:25 crc kubenswrapper[5035]: I1124 19:44:25.648201 5035 scope.go:117] "RemoveContainer" containerID="aaf5e478c74d73a669b311da730f60aedba30b4a400ba8647986584232a6ee5c" Nov 24 19:44:25 crc kubenswrapper[5035]: I1124 19:44:25.673470 5035 scope.go:117] "RemoveContainer" containerID="8b59b43ccd6101b25ddc3807830175ef03b74064e8932358c6f36b5441c3bde6" Nov 24 19:44:25 crc kubenswrapper[5035]: I1124 19:44:25.755627 5035 scope.go:117] "RemoveContainer" containerID="5c823d874fa8a07d068941d4574987a0b14864ec059594ce6afda6fc58ad9cf2" Nov 24 19:44:25 crc kubenswrapper[5035]: I1124 19:44:25.800697 5035 scope.go:117] "RemoveContainer" containerID="0db161e8e9b3e1ee285b374726011b12f15ed9a9b9ead3ef235b3170a01c0354" Nov 24 19:44:25 crc kubenswrapper[5035]: I1124 19:44:25.851139 5035 scope.go:117] "RemoveContainer" containerID="f4bc7e2ab83e1721cb4ceba99420481052776e7ed7c3515ce301ef7cf0cc6bb6" Nov 24 19:44:25 crc kubenswrapper[5035]: I1124 19:44:25.905271 5035 scope.go:117] "RemoveContainer" containerID="c7139be1d7916cf66d73e73bde008e314e9f05968ed9a5cdb16dd5962fe4855a" Nov 24 19:44:28 crc kubenswrapper[5035]: I1124 19:44:28.044582 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pwhl5"] Nov 24 19:44:28 crc kubenswrapper[5035]: I1124 19:44:28.052716 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pwhl5"] Nov 24 19:44:28 crc kubenswrapper[5035]: I1124 19:44:28.216731 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb" path="/var/lib/kubelet/pods/d77dcfaa-154f-4ed4-9fb3-abc2c6af70eb/volumes" Nov 24 19:44:36 crc kubenswrapper[5035]: I1124 19:44:36.486748 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:44:36 crc kubenswrapper[5035]: E1124 19:44:36.487477 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:44:45 crc kubenswrapper[5035]: I1124 19:44:45.044666 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xx7p9"] Nov 24 19:44:45 crc kubenswrapper[5035]: I1124 19:44:45.060208 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xx7p9"] Nov 24 19:44:46 crc kubenswrapper[5035]: I1124 19:44:46.025861 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lhgxj"] Nov 24 19:44:46 crc kubenswrapper[5035]: I1124 19:44:46.034683 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lhgxj"] Nov 24 19:44:46 crc kubenswrapper[5035]: I1124 19:44:46.213522 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46c96d7f-2de1-4104-9058-e9fb44e3b677" path="/var/lib/kubelet/pods/46c96d7f-2de1-4104-9058-e9fb44e3b677/volumes" Nov 24 19:44:46 crc kubenswrapper[5035]: I1124 19:44:46.214239 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf6870c-e93e-44db-9059-3ca4b23abab3" path="/var/lib/kubelet/pods/ebf6870c-e93e-44db-9059-3ca4b23abab3/volumes" Nov 24 19:44:47 crc kubenswrapper[5035]: I1124 19:44:47.200147 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:44:47 crc kubenswrapper[5035]: E1124 19:44:47.200683 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:44:50 crc kubenswrapper[5035]: I1124 19:44:50.950123 5035 generic.go:334] "Generic (PLEG): container finished" podID="945f0c1b-86e7-4759-96fb-cc5278b03215" containerID="1ae38fe2079d3cb2feb21ffbbe6deb6997d914223fdc20ece9b53910be21b0c3" exitCode=0 Nov 24 19:44:50 crc kubenswrapper[5035]: I1124 19:44:50.950193 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" event={"ID":"945f0c1b-86e7-4759-96fb-cc5278b03215","Type":"ContainerDied","Data":"1ae38fe2079d3cb2feb21ffbbe6deb6997d914223fdc20ece9b53910be21b0c3"} Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.403255 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.594156 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj4nd\" (UniqueName: \"kubernetes.io/projected/945f0c1b-86e7-4759-96fb-cc5278b03215-kube-api-access-gj4nd\") pod \"945f0c1b-86e7-4759-96fb-cc5278b03215\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.594564 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-ssh-key\") pod \"945f0c1b-86e7-4759-96fb-cc5278b03215\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.594834 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-inventory\") pod \"945f0c1b-86e7-4759-96fb-cc5278b03215\" (UID: \"945f0c1b-86e7-4759-96fb-cc5278b03215\") " Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.602325 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/945f0c1b-86e7-4759-96fb-cc5278b03215-kube-api-access-gj4nd" (OuterVolumeSpecName: "kube-api-access-gj4nd") pod "945f0c1b-86e7-4759-96fb-cc5278b03215" (UID: "945f0c1b-86e7-4759-96fb-cc5278b03215"). InnerVolumeSpecName "kube-api-access-gj4nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.641531 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-inventory" (OuterVolumeSpecName: "inventory") pod "945f0c1b-86e7-4759-96fb-cc5278b03215" (UID: "945f0c1b-86e7-4759-96fb-cc5278b03215"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.653137 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "945f0c1b-86e7-4759-96fb-cc5278b03215" (UID: "945f0c1b-86e7-4759-96fb-cc5278b03215"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.698132 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.698579 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj4nd\" (UniqueName: \"kubernetes.io/projected/945f0c1b-86e7-4759-96fb-cc5278b03215-kube-api-access-gj4nd\") on node \"crc\" DevicePath \"\"" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.699585 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/945f0c1b-86e7-4759-96fb-cc5278b03215-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.972527 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" event={"ID":"945f0c1b-86e7-4759-96fb-cc5278b03215","Type":"ContainerDied","Data":"add4b72a8fe607e58359a899a2397f87d11017c94b59c13f13c941750a9d3e4e"} Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.972582 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="add4b72a8fe607e58359a899a2397f87d11017c94b59c13f13c941750a9d3e4e" Nov 24 19:44:52 crc kubenswrapper[5035]: I1124 19:44:52.972612 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.065308 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4qf8z"] Nov 24 19:44:53 crc kubenswrapper[5035]: E1124 19:44:53.065742 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="945f0c1b-86e7-4759-96fb-cc5278b03215" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.065762 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="945f0c1b-86e7-4759-96fb-cc5278b03215" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.065978 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="945f0c1b-86e7-4759-96fb-cc5278b03215" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.066927 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.071754 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.071918 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.072021 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.072125 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.073273 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4qf8z"] Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.107688 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rmkx\" (UniqueName: \"kubernetes.io/projected/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-kube-api-access-7rmkx\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.107955 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.107980 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.209623 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rmkx\" (UniqueName: \"kubernetes.io/projected/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-kube-api-access-7rmkx\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.209669 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.209691 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.213755 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.216165 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.228654 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rmkx\" (UniqueName: \"kubernetes.io/projected/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-kube-api-access-7rmkx\") pod \"ssh-known-hosts-edpm-deployment-4qf8z\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.384957 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.939405 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4qf8z"] Nov 24 19:44:53 crc kubenswrapper[5035]: I1124 19:44:53.980841 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" event={"ID":"d74ff699-d2f6-4adc-b70a-b1a35e2d997a","Type":"ContainerStarted","Data":"0990bffe3e5fb9545f72fb73b5ecef89dac20d39f05a958b4c654c4197feb9e2"} Nov 24 19:44:54 crc kubenswrapper[5035]: I1124 19:44:54.990847 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" event={"ID":"d74ff699-d2f6-4adc-b70a-b1a35e2d997a","Type":"ContainerStarted","Data":"20ea6af7e5817c8b7c8cfa8ed5ce976d5e502040602f74bbf67cfd06056ed2f1"} Nov 24 19:44:55 crc kubenswrapper[5035]: I1124 19:44:55.013489 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" podStartSLOduration=1.513828634 podStartE2EDuration="2.013472742s" podCreationTimestamp="2025-11-24 19:44:53 +0000 UTC" firstStartedPulling="2025-11-24 19:44:53.942144642 +0000 UTC m=+1832.464650899" lastFinishedPulling="2025-11-24 19:44:54.44178874 +0000 UTC m=+1832.964295007" observedRunningTime="2025-11-24 19:44:55.007912531 +0000 UTC m=+1833.530418788" watchObservedRunningTime="2025-11-24 19:44:55.013472742 +0000 UTC m=+1833.535978999" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.163268 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx"] Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.165728 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.168053 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.168330 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.178622 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx"] Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.243626 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv64n\" (UniqueName: \"kubernetes.io/projected/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-kube-api-access-rv64n\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.243669 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-secret-volume\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.244019 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-config-volume\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.345611 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-secret-volume\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.345755 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-config-volume\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.345872 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv64n\" (UniqueName: \"kubernetes.io/projected/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-kube-api-access-rv64n\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.347068 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-config-volume\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.352766 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-secret-volume\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.365578 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv64n\" (UniqueName: \"kubernetes.io/projected/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-kube-api-access-rv64n\") pod \"collect-profiles-29400225-v6cvx\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.495786 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:00 crc kubenswrapper[5035]: I1124 19:45:00.940416 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx"] Nov 24 19:45:01 crc kubenswrapper[5035]: I1124 19:45:01.066044 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" event={"ID":"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7","Type":"ContainerStarted","Data":"161e72e9cd3765f3c22714e1c3c305fb98b37bcf6daed6c7e686481fc06e2d35"} Nov 24 19:45:01 crc kubenswrapper[5035]: I1124 19:45:01.200324 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:45:01 crc kubenswrapper[5035]: E1124 19:45:01.200693 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:45:02 crc kubenswrapper[5035]: I1124 19:45:02.075735 5035 generic.go:334] "Generic (PLEG): container finished" podID="d74ff699-d2f6-4adc-b70a-b1a35e2d997a" containerID="20ea6af7e5817c8b7c8cfa8ed5ce976d5e502040602f74bbf67cfd06056ed2f1" exitCode=0 Nov 24 19:45:02 crc kubenswrapper[5035]: I1124 19:45:02.075811 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" event={"ID":"d74ff699-d2f6-4adc-b70a-b1a35e2d997a","Type":"ContainerDied","Data":"20ea6af7e5817c8b7c8cfa8ed5ce976d5e502040602f74bbf67cfd06056ed2f1"} Nov 24 19:45:02 crc kubenswrapper[5035]: I1124 19:45:02.078183 5035 generic.go:334] "Generic (PLEG): container finished" podID="3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" containerID="298a61232d0326191769c34ceede8538573bdc760d0a8398c6cc9760d7f349a3" exitCode=0 Nov 24 19:45:02 crc kubenswrapper[5035]: I1124 19:45:02.078257 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" event={"ID":"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7","Type":"ContainerDied","Data":"298a61232d0326191769c34ceede8538573bdc760d0a8398c6cc9760d7f349a3"} Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.521762 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.526281 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.712010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-config-volume\") pod \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.712059 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv64n\" (UniqueName: \"kubernetes.io/projected/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-kube-api-access-rv64n\") pod \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.712184 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-secret-volume\") pod \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\" (UID: \"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7\") " Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.712244 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-inventory-0\") pod \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.712376 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-ssh-key-openstack-edpm-ipam\") pod \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.712429 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rmkx\" (UniqueName: \"kubernetes.io/projected/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-kube-api-access-7rmkx\") pod \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\" (UID: \"d74ff699-d2f6-4adc-b70a-b1a35e2d997a\") " Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.713006 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-config-volume" (OuterVolumeSpecName: "config-volume") pod "3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" (UID: "3b2065a7-34e8-4aa0-ab18-60c002c5fbc7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.720984 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" (UID: "3b2065a7-34e8-4aa0-ab18-60c002c5fbc7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.721556 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-kube-api-access-rv64n" (OuterVolumeSpecName: "kube-api-access-rv64n") pod "3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" (UID: "3b2065a7-34e8-4aa0-ab18-60c002c5fbc7"). InnerVolumeSpecName "kube-api-access-rv64n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.722496 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-kube-api-access-7rmkx" (OuterVolumeSpecName: "kube-api-access-7rmkx") pod "d74ff699-d2f6-4adc-b70a-b1a35e2d997a" (UID: "d74ff699-d2f6-4adc-b70a-b1a35e2d997a"). InnerVolumeSpecName "kube-api-access-7rmkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.739379 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d74ff699-d2f6-4adc-b70a-b1a35e2d997a" (UID: "d74ff699-d2f6-4adc-b70a-b1a35e2d997a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.749512 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d74ff699-d2f6-4adc-b70a-b1a35e2d997a" (UID: "d74ff699-d2f6-4adc-b70a-b1a35e2d997a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.814546 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.814587 5035 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.814601 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.814615 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rmkx\" (UniqueName: \"kubernetes.io/projected/d74ff699-d2f6-4adc-b70a-b1a35e2d997a-kube-api-access-7rmkx\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.814626 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:03 crc kubenswrapper[5035]: I1124 19:45:03.814638 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv64n\" (UniqueName: \"kubernetes.io/projected/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7-kube-api-access-rv64n\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.100900 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" event={"ID":"d74ff699-d2f6-4adc-b70a-b1a35e2d997a","Type":"ContainerDied","Data":"0990bffe3e5fb9545f72fb73b5ecef89dac20d39f05a958b4c654c4197feb9e2"} Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.100968 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0990bffe3e5fb9545f72fb73b5ecef89dac20d39f05a958b4c654c4197feb9e2" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.101450 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4qf8z" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.102323 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" event={"ID":"3b2065a7-34e8-4aa0-ab18-60c002c5fbc7","Type":"ContainerDied","Data":"161e72e9cd3765f3c22714e1c3c305fb98b37bcf6daed6c7e686481fc06e2d35"} Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.102371 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="161e72e9cd3765f3c22714e1c3c305fb98b37bcf6daed6c7e686481fc06e2d35" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.102392 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.187338 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7"] Nov 24 19:45:04 crc kubenswrapper[5035]: E1124 19:45:04.187805 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d74ff699-d2f6-4adc-b70a-b1a35e2d997a" containerName="ssh-known-hosts-edpm-deployment" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.187827 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d74ff699-d2f6-4adc-b70a-b1a35e2d997a" containerName="ssh-known-hosts-edpm-deployment" Nov 24 19:45:04 crc kubenswrapper[5035]: E1124 19:45:04.187862 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" containerName="collect-profiles" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.187872 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" containerName="collect-profiles" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.188071 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" containerName="collect-profiles" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.188104 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d74ff699-d2f6-4adc-b70a-b1a35e2d997a" containerName="ssh-known-hosts-edpm-deployment" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.188803 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.199139 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.200055 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.200508 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.201213 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.227882 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7"] Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.328014 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqw22\" (UniqueName: \"kubernetes.io/projected/a7ef3873-f86b-41de-9a54-96e8c47e5e16-kube-api-access-wqw22\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.328132 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.328183 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.429243 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.429324 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.429458 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqw22\" (UniqueName: \"kubernetes.io/projected/a7ef3873-f86b-41de-9a54-96e8c47e5e16-kube-api-access-wqw22\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.435335 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.437458 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.456659 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqw22\" (UniqueName: \"kubernetes.io/projected/a7ef3873-f86b-41de-9a54-96e8c47e5e16-kube-api-access-wqw22\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9ngj7\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:04 crc kubenswrapper[5035]: I1124 19:45:04.518456 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:05 crc kubenswrapper[5035]: I1124 19:45:05.039754 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7"] Nov 24 19:45:05 crc kubenswrapper[5035]: I1124 19:45:05.113210 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" event={"ID":"a7ef3873-f86b-41de-9a54-96e8c47e5e16","Type":"ContainerStarted","Data":"8bdd107cf3c56eb4246bfefcc0005a3bab8756d26852264979d733c878a4c1b3"} Nov 24 19:45:06 crc kubenswrapper[5035]: I1124 19:45:06.122444 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" event={"ID":"a7ef3873-f86b-41de-9a54-96e8c47e5e16","Type":"ContainerStarted","Data":"45a09b5f169447bad1fd6d005942767fc7cefb67e285cf436876df57268d0278"} Nov 24 19:45:06 crc kubenswrapper[5035]: I1124 19:45:06.141830 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" podStartSLOduration=1.551832461 podStartE2EDuration="2.141812731s" podCreationTimestamp="2025-11-24 19:45:04 +0000 UTC" firstStartedPulling="2025-11-24 19:45:05.043918818 +0000 UTC m=+1843.566425075" lastFinishedPulling="2025-11-24 19:45:05.633899058 +0000 UTC m=+1844.156405345" observedRunningTime="2025-11-24 19:45:06.13847614 +0000 UTC m=+1844.660982397" watchObservedRunningTime="2025-11-24 19:45:06.141812731 +0000 UTC m=+1844.664318988" Nov 24 19:45:13 crc kubenswrapper[5035]: I1124 19:45:13.200348 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:45:13 crc kubenswrapper[5035]: E1124 19:45:13.201362 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:45:15 crc kubenswrapper[5035]: I1124 19:45:15.220235 5035 generic.go:334] "Generic (PLEG): container finished" podID="a7ef3873-f86b-41de-9a54-96e8c47e5e16" containerID="45a09b5f169447bad1fd6d005942767fc7cefb67e285cf436876df57268d0278" exitCode=0 Nov 24 19:45:15 crc kubenswrapper[5035]: I1124 19:45:15.220404 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" event={"ID":"a7ef3873-f86b-41de-9a54-96e8c47e5e16","Type":"ContainerDied","Data":"45a09b5f169447bad1fd6d005942767fc7cefb67e285cf436876df57268d0278"} Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.616659 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.674535 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqw22\" (UniqueName: \"kubernetes.io/projected/a7ef3873-f86b-41de-9a54-96e8c47e5e16-kube-api-access-wqw22\") pod \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.674954 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-inventory\") pod \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.675663 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-ssh-key\") pod \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\" (UID: \"a7ef3873-f86b-41de-9a54-96e8c47e5e16\") " Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.680799 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ef3873-f86b-41de-9a54-96e8c47e5e16-kube-api-access-wqw22" (OuterVolumeSpecName: "kube-api-access-wqw22") pod "a7ef3873-f86b-41de-9a54-96e8c47e5e16" (UID: "a7ef3873-f86b-41de-9a54-96e8c47e5e16"). InnerVolumeSpecName "kube-api-access-wqw22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.722006 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a7ef3873-f86b-41de-9a54-96e8c47e5e16" (UID: "a7ef3873-f86b-41de-9a54-96e8c47e5e16"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.731219 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-inventory" (OuterVolumeSpecName: "inventory") pod "a7ef3873-f86b-41de-9a54-96e8c47e5e16" (UID: "a7ef3873-f86b-41de-9a54-96e8c47e5e16"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.778172 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqw22\" (UniqueName: \"kubernetes.io/projected/a7ef3873-f86b-41de-9a54-96e8c47e5e16-kube-api-access-wqw22\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.778224 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:16 crc kubenswrapper[5035]: I1124 19:45:16.778240 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7ef3873-f86b-41de-9a54-96e8c47e5e16-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.240512 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" event={"ID":"a7ef3873-f86b-41de-9a54-96e8c47e5e16","Type":"ContainerDied","Data":"8bdd107cf3c56eb4246bfefcc0005a3bab8756d26852264979d733c878a4c1b3"} Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.240562 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bdd107cf3c56eb4246bfefcc0005a3bab8756d26852264979d733c878a4c1b3" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.240585 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.334273 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq"] Nov 24 19:45:17 crc kubenswrapper[5035]: E1124 19:45:17.334774 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ef3873-f86b-41de-9a54-96e8c47e5e16" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.334799 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ef3873-f86b-41de-9a54-96e8c47e5e16" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.335017 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ef3873-f86b-41de-9a54-96e8c47e5e16" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.335879 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.340352 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.340372 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.340620 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.341207 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.360437 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq"] Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.409818 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn22d\" (UniqueName: \"kubernetes.io/projected/d3eadfa2-fc9a-49af-8e88-1c83793f075a-kube-api-access-gn22d\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.409913 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.410005 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.511923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.512049 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.512109 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn22d\" (UniqueName: \"kubernetes.io/projected/d3eadfa2-fc9a-49af-8e88-1c83793f075a-kube-api-access-gn22d\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.519442 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.519440 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.529371 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn22d\" (UniqueName: \"kubernetes.io/projected/d3eadfa2-fc9a-49af-8e88-1c83793f075a-kube-api-access-gn22d\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:17 crc kubenswrapper[5035]: I1124 19:45:17.658031 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:18 crc kubenswrapper[5035]: I1124 19:45:18.178506 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq"] Nov 24 19:45:18 crc kubenswrapper[5035]: I1124 19:45:18.248376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" event={"ID":"d3eadfa2-fc9a-49af-8e88-1c83793f075a","Type":"ContainerStarted","Data":"da6ff2fea50d7977047277c3051ab1a132c7b2bbad76f0abc309dee1d89e2908"} Nov 24 19:45:19 crc kubenswrapper[5035]: I1124 19:45:19.264751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" event={"ID":"d3eadfa2-fc9a-49af-8e88-1c83793f075a","Type":"ContainerStarted","Data":"3f336398724b57b54db8dcd327d90f55c7184faec5ac1bd3368732acda2caca4"} Nov 24 19:45:19 crc kubenswrapper[5035]: I1124 19:45:19.295856 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" podStartSLOduration=1.873558317 podStartE2EDuration="2.295830846s" podCreationTimestamp="2025-11-24 19:45:17 +0000 UTC" firstStartedPulling="2025-11-24 19:45:18.181955988 +0000 UTC m=+1856.704462245" lastFinishedPulling="2025-11-24 19:45:18.604228507 +0000 UTC m=+1857.126734774" observedRunningTime="2025-11-24 19:45:19.284052676 +0000 UTC m=+1857.806558933" watchObservedRunningTime="2025-11-24 19:45:19.295830846 +0000 UTC m=+1857.818337123" Nov 24 19:45:24 crc kubenswrapper[5035]: I1124 19:45:24.201368 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:45:25 crc kubenswrapper[5035]: I1124 19:45:25.324328 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"614af86f0ff3c54135bc9d250592e150a845656f46635f75ac952947e7986d0f"} Nov 24 19:45:26 crc kubenswrapper[5035]: I1124 19:45:26.077052 5035 scope.go:117] "RemoveContainer" containerID="dea16e9e471c63ab5f4ee56d465ac26464ba9db13439aed60667c20c0f50a9aa" Nov 24 19:45:26 crc kubenswrapper[5035]: I1124 19:45:26.129492 5035 scope.go:117] "RemoveContainer" containerID="65b044e2ccc2eb9a95bc646b33ae055592bc0466d9ec4a29a3eb320809c90321" Nov 24 19:45:26 crc kubenswrapper[5035]: I1124 19:45:26.191908 5035 scope.go:117] "RemoveContainer" containerID="b5223d0a74f3ad3936dff3f303b7933a624ec0175ef8185d0bd50636ec2e1723" Nov 24 19:45:26 crc kubenswrapper[5035]: I1124 19:45:26.217468 5035 scope.go:117] "RemoveContainer" containerID="1f36c22ea2fcdb76e4f77e39a76ddda86e0b2ca386add147433a833e2949e68e" Nov 24 19:45:26 crc kubenswrapper[5035]: I1124 19:45:26.240988 5035 scope.go:117] "RemoveContainer" containerID="2509916e3fa226dd2f630b71b795266388a2c1a13c67d492fe0fc77b89e04125" Nov 24 19:45:26 crc kubenswrapper[5035]: I1124 19:45:26.300146 5035 scope.go:117] "RemoveContainer" containerID="e56bf8c1e6cddfddbed2480e5692e72a825a250a8e1cf9e685bd2a77bd373916" Nov 24 19:45:29 crc kubenswrapper[5035]: I1124 19:45:29.361363 5035 generic.go:334] "Generic (PLEG): container finished" podID="d3eadfa2-fc9a-49af-8e88-1c83793f075a" containerID="3f336398724b57b54db8dcd327d90f55c7184faec5ac1bd3368732acda2caca4" exitCode=0 Nov 24 19:45:29 crc kubenswrapper[5035]: I1124 19:45:29.361435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" event={"ID":"d3eadfa2-fc9a-49af-8e88-1c83793f075a","Type":"ContainerDied","Data":"3f336398724b57b54db8dcd327d90f55c7184faec5ac1bd3368732acda2caca4"} Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.066849 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-gnkh9"] Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.081414 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-gnkh9"] Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.223363 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e36d14-3aed-4554-8cab-f20fa21a29d8" path="/var/lib/kubelet/pods/99e36d14-3aed-4554-8cab-f20fa21a29d8/volumes" Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.857612 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.918547 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-inventory\") pod \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.920101 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-ssh-key\") pod \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.920253 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn22d\" (UniqueName: \"kubernetes.io/projected/d3eadfa2-fc9a-49af-8e88-1c83793f075a-kube-api-access-gn22d\") pod \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\" (UID: \"d3eadfa2-fc9a-49af-8e88-1c83793f075a\") " Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.925772 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3eadfa2-fc9a-49af-8e88-1c83793f075a-kube-api-access-gn22d" (OuterVolumeSpecName: "kube-api-access-gn22d") pod "d3eadfa2-fc9a-49af-8e88-1c83793f075a" (UID: "d3eadfa2-fc9a-49af-8e88-1c83793f075a"). InnerVolumeSpecName "kube-api-access-gn22d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.948925 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d3eadfa2-fc9a-49af-8e88-1c83793f075a" (UID: "d3eadfa2-fc9a-49af-8e88-1c83793f075a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:45:30 crc kubenswrapper[5035]: I1124 19:45:30.948975 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-inventory" (OuterVolumeSpecName: "inventory") pod "d3eadfa2-fc9a-49af-8e88-1c83793f075a" (UID: "d3eadfa2-fc9a-49af-8e88-1c83793f075a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:45:31 crc kubenswrapper[5035]: I1124 19:45:31.022876 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn22d\" (UniqueName: \"kubernetes.io/projected/d3eadfa2-fc9a-49af-8e88-1c83793f075a-kube-api-access-gn22d\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:31 crc kubenswrapper[5035]: I1124 19:45:31.022904 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:31 crc kubenswrapper[5035]: I1124 19:45:31.022914 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3eadfa2-fc9a-49af-8e88-1c83793f075a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:45:31 crc kubenswrapper[5035]: I1124 19:45:31.386706 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" event={"ID":"d3eadfa2-fc9a-49af-8e88-1c83793f075a","Type":"ContainerDied","Data":"da6ff2fea50d7977047277c3051ab1a132c7b2bbad76f0abc309dee1d89e2908"} Nov 24 19:45:31 crc kubenswrapper[5035]: I1124 19:45:31.386762 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da6ff2fea50d7977047277c3051ab1a132c7b2bbad76f0abc309dee1d89e2908" Nov 24 19:45:31 crc kubenswrapper[5035]: I1124 19:45:31.386832 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq" Nov 24 19:46:26 crc kubenswrapper[5035]: I1124 19:46:26.383900 5035 scope.go:117] "RemoveContainer" containerID="b2a4c21951ea187c9e40b0b6d4b3ce1c29e72c640188b0c8eeed8acf139c21ce" Nov 24 19:47:45 crc kubenswrapper[5035]: I1124 19:47:45.234034 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:47:45 crc kubenswrapper[5035]: I1124 19:47:45.235488 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:48:15 crc kubenswrapper[5035]: I1124 19:48:15.234922 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:48:15 crc kubenswrapper[5035]: I1124 19:48:15.235728 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.180495 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6p2s7"] Nov 24 19:48:39 crc kubenswrapper[5035]: E1124 19:48:39.182744 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3eadfa2-fc9a-49af-8e88-1c83793f075a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.182780 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3eadfa2-fc9a-49af-8e88-1c83793f075a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.183282 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3eadfa2-fc9a-49af-8e88-1c83793f075a" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.185603 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.203658 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6p2s7"] Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.224536 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-utilities\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.224902 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-catalog-content\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.225265 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92mv7\" (UniqueName: \"kubernetes.io/projected/d40054ec-0c14-4934-8ab0-0c26350f6f3b-kube-api-access-92mv7\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.327167 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92mv7\" (UniqueName: \"kubernetes.io/projected/d40054ec-0c14-4934-8ab0-0c26350f6f3b-kube-api-access-92mv7\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.327270 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-utilities\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.327314 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-catalog-content\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.327899 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-utilities\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.327958 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-catalog-content\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.346141 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92mv7\" (UniqueName: \"kubernetes.io/projected/d40054ec-0c14-4934-8ab0-0c26350f6f3b-kube-api-access-92mv7\") pod \"redhat-operators-6p2s7\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.526500 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:39 crc kubenswrapper[5035]: I1124 19:48:39.790829 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6p2s7"] Nov 24 19:48:40 crc kubenswrapper[5035]: I1124 19:48:40.518075 5035 generic.go:334] "Generic (PLEG): container finished" podID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerID="b47f815dd6398b01522c3e23debee4d7bb4fbbe0ab0bb8ca6da42783d8a3b553" exitCode=0 Nov 24 19:48:40 crc kubenswrapper[5035]: I1124 19:48:40.518142 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p2s7" event={"ID":"d40054ec-0c14-4934-8ab0-0c26350f6f3b","Type":"ContainerDied","Data":"b47f815dd6398b01522c3e23debee4d7bb4fbbe0ab0bb8ca6da42783d8a3b553"} Nov 24 19:48:40 crc kubenswrapper[5035]: I1124 19:48:40.518395 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p2s7" event={"ID":"d40054ec-0c14-4934-8ab0-0c26350f6f3b","Type":"ContainerStarted","Data":"6e7cf63d4fb7e9e17542fe9898db1bbc64f17d18b0e582e79b53347c5ffad627"} Nov 24 19:48:40 crc kubenswrapper[5035]: I1124 19:48:40.520034 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 19:48:42 crc kubenswrapper[5035]: I1124 19:48:42.538707 5035 generic.go:334] "Generic (PLEG): container finished" podID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerID="3aa8eaa6474bd50647140df0f3a66efdeaf0cc964b8d42e996c86cbcd248a3ea" exitCode=0 Nov 24 19:48:42 crc kubenswrapper[5035]: I1124 19:48:42.538843 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p2s7" event={"ID":"d40054ec-0c14-4934-8ab0-0c26350f6f3b","Type":"ContainerDied","Data":"3aa8eaa6474bd50647140df0f3a66efdeaf0cc964b8d42e996c86cbcd248a3ea"} Nov 24 19:48:43 crc kubenswrapper[5035]: I1124 19:48:43.550869 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p2s7" event={"ID":"d40054ec-0c14-4934-8ab0-0c26350f6f3b","Type":"ContainerStarted","Data":"35da153c753527e054afed306e7e32ccfe938a1f992325fcf78795798b00321f"} Nov 24 19:48:43 crc kubenswrapper[5035]: I1124 19:48:43.571791 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6p2s7" podStartSLOduration=1.806279531 podStartE2EDuration="4.571770819s" podCreationTimestamp="2025-11-24 19:48:39 +0000 UTC" firstStartedPulling="2025-11-24 19:48:40.519781132 +0000 UTC m=+2059.042287389" lastFinishedPulling="2025-11-24 19:48:43.28527242 +0000 UTC m=+2061.807778677" observedRunningTime="2025-11-24 19:48:43.56775044 +0000 UTC m=+2062.090256737" watchObservedRunningTime="2025-11-24 19:48:43.571770819 +0000 UTC m=+2062.094277086" Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.235088 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.235191 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.235273 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.236578 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"614af86f0ff3c54135bc9d250592e150a845656f46635f75ac952947e7986d0f"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.236712 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://614af86f0ff3c54135bc9d250592e150a845656f46635f75ac952947e7986d0f" gracePeriod=600 Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.571954 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="614af86f0ff3c54135bc9d250592e150a845656f46635f75ac952947e7986d0f" exitCode=0 Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.572141 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"614af86f0ff3c54135bc9d250592e150a845656f46635f75ac952947e7986d0f"} Nov 24 19:48:45 crc kubenswrapper[5035]: I1124 19:48:45.572726 5035 scope.go:117] "RemoveContainer" containerID="e07c791ae07a90c073ac624c179222ef511c97936e049c20774026edbf733199" Nov 24 19:48:46 crc kubenswrapper[5035]: I1124 19:48:46.586750 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20"} Nov 24 19:48:49 crc kubenswrapper[5035]: I1124 19:48:49.527091 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:49 crc kubenswrapper[5035]: I1124 19:48:49.527742 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:49 crc kubenswrapper[5035]: I1124 19:48:49.580034 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:49 crc kubenswrapper[5035]: I1124 19:48:49.678565 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:49 crc kubenswrapper[5035]: I1124 19:48:49.812338 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6p2s7"] Nov 24 19:48:51 crc kubenswrapper[5035]: I1124 19:48:51.653460 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6p2s7" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="registry-server" containerID="cri-o://35da153c753527e054afed306e7e32ccfe938a1f992325fcf78795798b00321f" gracePeriod=2 Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.691970 5035 generic.go:334] "Generic (PLEG): container finished" podID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerID="35da153c753527e054afed306e7e32ccfe938a1f992325fcf78795798b00321f" exitCode=0 Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.692802 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p2s7" event={"ID":"d40054ec-0c14-4934-8ab0-0c26350f6f3b","Type":"ContainerDied","Data":"35da153c753527e054afed306e7e32ccfe938a1f992325fcf78795798b00321f"} Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.794514 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.938196 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-catalog-content\") pod \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.938486 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92mv7\" (UniqueName: \"kubernetes.io/projected/d40054ec-0c14-4934-8ab0-0c26350f6f3b-kube-api-access-92mv7\") pod \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.938757 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-utilities\") pod \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\" (UID: \"d40054ec-0c14-4934-8ab0-0c26350f6f3b\") " Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.940355 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-utilities" (OuterVolumeSpecName: "utilities") pod "d40054ec-0c14-4934-8ab0-0c26350f6f3b" (UID: "d40054ec-0c14-4934-8ab0-0c26350f6f3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:48:54 crc kubenswrapper[5035]: I1124 19:48:54.945000 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40054ec-0c14-4934-8ab0-0c26350f6f3b-kube-api-access-92mv7" (OuterVolumeSpecName: "kube-api-access-92mv7") pod "d40054ec-0c14-4934-8ab0-0c26350f6f3b" (UID: "d40054ec-0c14-4934-8ab0-0c26350f6f3b"). InnerVolumeSpecName "kube-api-access-92mv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.025027 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d40054ec-0c14-4934-8ab0-0c26350f6f3b" (UID: "d40054ec-0c14-4934-8ab0-0c26350f6f3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.041520 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.041551 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92mv7\" (UniqueName: \"kubernetes.io/projected/d40054ec-0c14-4934-8ab0-0c26350f6f3b-kube-api-access-92mv7\") on node \"crc\" DevicePath \"\"" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.041564 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40054ec-0c14-4934-8ab0-0c26350f6f3b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.704139 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p2s7" event={"ID":"d40054ec-0c14-4934-8ab0-0c26350f6f3b","Type":"ContainerDied","Data":"6e7cf63d4fb7e9e17542fe9898db1bbc64f17d18b0e582e79b53347c5ffad627"} Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.704195 5035 scope.go:117] "RemoveContainer" containerID="35da153c753527e054afed306e7e32ccfe938a1f992325fcf78795798b00321f" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.704947 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p2s7" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.733947 5035 scope.go:117] "RemoveContainer" containerID="3aa8eaa6474bd50647140df0f3a66efdeaf0cc964b8d42e996c86cbcd248a3ea" Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.740660 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6p2s7"] Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.746224 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6p2s7"] Nov 24 19:48:55 crc kubenswrapper[5035]: I1124 19:48:55.760346 5035 scope.go:117] "RemoveContainer" containerID="b47f815dd6398b01522c3e23debee4d7bb4fbbe0ab0bb8ca6da42783d8a3b553" Nov 24 19:48:56 crc kubenswrapper[5035]: I1124 19:48:56.211768 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" path="/var/lib/kubelet/pods/d40054ec-0c14-4934-8ab0-0c26350f6f3b/volumes" Nov 24 19:49:16 crc kubenswrapper[5035]: I1124 19:49:16.992052 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n76jt"] Nov 24 19:49:16 crc kubenswrapper[5035]: E1124 19:49:16.993014 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="registry-server" Nov 24 19:49:16 crc kubenswrapper[5035]: I1124 19:49:16.993031 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="registry-server" Nov 24 19:49:16 crc kubenswrapper[5035]: E1124 19:49:16.993069 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="extract-utilities" Nov 24 19:49:16 crc kubenswrapper[5035]: I1124 19:49:16.993078 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="extract-utilities" Nov 24 19:49:16 crc kubenswrapper[5035]: E1124 19:49:16.993092 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="extract-content" Nov 24 19:49:16 crc kubenswrapper[5035]: I1124 19:49:16.993100 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="extract-content" Nov 24 19:49:16 crc kubenswrapper[5035]: I1124 19:49:16.993318 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40054ec-0c14-4934-8ab0-0c26350f6f3b" containerName="registry-server" Nov 24 19:49:16 crc kubenswrapper[5035]: I1124 19:49:16.995017 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.035636 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n76jt"] Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.060354 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhdv9\" (UniqueName: \"kubernetes.io/projected/cbabe1be-acac-4c15-9465-df84149a2dde-kube-api-access-hhdv9\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.060649 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-catalog-content\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.061121 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-utilities\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.162270 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhdv9\" (UniqueName: \"kubernetes.io/projected/cbabe1be-acac-4c15-9465-df84149a2dde-kube-api-access-hhdv9\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.162591 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-catalog-content\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.162745 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-utilities\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.163218 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-catalog-content\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.163227 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-utilities\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.189982 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhdv9\" (UniqueName: \"kubernetes.io/projected/cbabe1be-acac-4c15-9465-df84149a2dde-kube-api-access-hhdv9\") pod \"certified-operators-n76jt\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.336323 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:17 crc kubenswrapper[5035]: I1124 19:49:17.896026 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n76jt"] Nov 24 19:49:18 crc kubenswrapper[5035]: I1124 19:49:18.901994 5035 generic.go:334] "Generic (PLEG): container finished" podID="cbabe1be-acac-4c15-9465-df84149a2dde" containerID="eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd" exitCode=0 Nov 24 19:49:18 crc kubenswrapper[5035]: I1124 19:49:18.902060 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n76jt" event={"ID":"cbabe1be-acac-4c15-9465-df84149a2dde","Type":"ContainerDied","Data":"eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd"} Nov 24 19:49:18 crc kubenswrapper[5035]: I1124 19:49:18.902261 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n76jt" event={"ID":"cbabe1be-acac-4c15-9465-df84149a2dde","Type":"ContainerStarted","Data":"a87594fa6051b90013990f888600f7a35345c8f33a4011ccbf8a56f14f367bd0"} Nov 24 19:49:19 crc kubenswrapper[5035]: I1124 19:49:19.914213 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n76jt" event={"ID":"cbabe1be-acac-4c15-9465-df84149a2dde","Type":"ContainerStarted","Data":"8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f"} Nov 24 19:49:20 crc kubenswrapper[5035]: I1124 19:49:20.928117 5035 generic.go:334] "Generic (PLEG): container finished" podID="cbabe1be-acac-4c15-9465-df84149a2dde" containerID="8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f" exitCode=0 Nov 24 19:49:20 crc kubenswrapper[5035]: I1124 19:49:20.928168 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n76jt" event={"ID":"cbabe1be-acac-4c15-9465-df84149a2dde","Type":"ContainerDied","Data":"8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f"} Nov 24 19:49:21 crc kubenswrapper[5035]: I1124 19:49:21.937243 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n76jt" event={"ID":"cbabe1be-acac-4c15-9465-df84149a2dde","Type":"ContainerStarted","Data":"bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6"} Nov 24 19:49:21 crc kubenswrapper[5035]: I1124 19:49:21.962943 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n76jt" podStartSLOduration=3.489212933 podStartE2EDuration="5.962917949s" podCreationTimestamp="2025-11-24 19:49:16 +0000 UTC" firstStartedPulling="2025-11-24 19:49:18.904493368 +0000 UTC m=+2097.426999635" lastFinishedPulling="2025-11-24 19:49:21.378198364 +0000 UTC m=+2099.900704651" observedRunningTime="2025-11-24 19:49:21.954520652 +0000 UTC m=+2100.477026909" watchObservedRunningTime="2025-11-24 19:49:21.962917949 +0000 UTC m=+2100.485424226" Nov 24 19:49:27 crc kubenswrapper[5035]: I1124 19:49:27.336402 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:27 crc kubenswrapper[5035]: I1124 19:49:27.337515 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:27 crc kubenswrapper[5035]: I1124 19:49:27.407406 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:28 crc kubenswrapper[5035]: I1124 19:49:28.055890 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:28 crc kubenswrapper[5035]: I1124 19:49:28.110653 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n76jt"] Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.015365 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n76jt" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="registry-server" containerID="cri-o://bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6" gracePeriod=2 Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.478804 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.628162 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-utilities\") pod \"cbabe1be-acac-4c15-9465-df84149a2dde\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.628248 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhdv9\" (UniqueName: \"kubernetes.io/projected/cbabe1be-acac-4c15-9465-df84149a2dde-kube-api-access-hhdv9\") pod \"cbabe1be-acac-4c15-9465-df84149a2dde\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.628455 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-catalog-content\") pod \"cbabe1be-acac-4c15-9465-df84149a2dde\" (UID: \"cbabe1be-acac-4c15-9465-df84149a2dde\") " Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.629264 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-utilities" (OuterVolumeSpecName: "utilities") pod "cbabe1be-acac-4c15-9465-df84149a2dde" (UID: "cbabe1be-acac-4c15-9465-df84149a2dde"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.636990 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.643671 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbabe1be-acac-4c15-9465-df84149a2dde-kube-api-access-hhdv9" (OuterVolumeSpecName: "kube-api-access-hhdv9") pod "cbabe1be-acac-4c15-9465-df84149a2dde" (UID: "cbabe1be-acac-4c15-9465-df84149a2dde"). InnerVolumeSpecName "kube-api-access-hhdv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.685158 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbabe1be-acac-4c15-9465-df84149a2dde" (UID: "cbabe1be-acac-4c15-9465-df84149a2dde"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.737956 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbabe1be-acac-4c15-9465-df84149a2dde-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:49:30 crc kubenswrapper[5035]: I1124 19:49:30.737995 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhdv9\" (UniqueName: \"kubernetes.io/projected/cbabe1be-acac-4c15-9465-df84149a2dde-kube-api-access-hhdv9\") on node \"crc\" DevicePath \"\"" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.026570 5035 generic.go:334] "Generic (PLEG): container finished" podID="cbabe1be-acac-4c15-9465-df84149a2dde" containerID="bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6" exitCode=0 Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.026655 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n76jt" event={"ID":"cbabe1be-acac-4c15-9465-df84149a2dde","Type":"ContainerDied","Data":"bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6"} Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.026798 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n76jt" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.028159 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n76jt" event={"ID":"cbabe1be-acac-4c15-9465-df84149a2dde","Type":"ContainerDied","Data":"a87594fa6051b90013990f888600f7a35345c8f33a4011ccbf8a56f14f367bd0"} Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.028245 5035 scope.go:117] "RemoveContainer" containerID="bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.068611 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n76jt"] Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.069163 5035 scope.go:117] "RemoveContainer" containerID="8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.075739 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n76jt"] Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.102105 5035 scope.go:117] "RemoveContainer" containerID="eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.133795 5035 scope.go:117] "RemoveContainer" containerID="bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6" Nov 24 19:49:31 crc kubenswrapper[5035]: E1124 19:49:31.134150 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6\": container with ID starting with bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6 not found: ID does not exist" containerID="bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.134190 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6"} err="failed to get container status \"bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6\": rpc error: code = NotFound desc = could not find container \"bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6\": container with ID starting with bb407c67e4276a95b0a9b10d4c1f42db28c7db8990d058848f4f8a3472671db6 not found: ID does not exist" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.134216 5035 scope.go:117] "RemoveContainer" containerID="8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f" Nov 24 19:49:31 crc kubenswrapper[5035]: E1124 19:49:31.134473 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f\": container with ID starting with 8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f not found: ID does not exist" containerID="8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.134498 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f"} err="failed to get container status \"8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f\": rpc error: code = NotFound desc = could not find container \"8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f\": container with ID starting with 8d1fe286cffa932118333296e27996c7974c107c8f1e1e0b812bb99b04038d2f not found: ID does not exist" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.134511 5035 scope.go:117] "RemoveContainer" containerID="eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd" Nov 24 19:49:31 crc kubenswrapper[5035]: E1124 19:49:31.134762 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd\": container with ID starting with eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd not found: ID does not exist" containerID="eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd" Nov 24 19:49:31 crc kubenswrapper[5035]: I1124 19:49:31.134811 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd"} err="failed to get container status \"eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd\": rpc error: code = NotFound desc = could not find container \"eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd\": container with ID starting with eb020467b5d047833b07b23a862593e1636bd315f621fc8acf797aa9751daabd not found: ID does not exist" Nov 24 19:49:32 crc kubenswrapper[5035]: I1124 19:49:32.213890 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" path="/var/lib/kubelet/pods/cbabe1be-acac-4c15-9465-df84149a2dde/volumes" Nov 24 19:49:42 crc kubenswrapper[5035]: E1124 19:49:42.230639 5035 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.47:58720->38.102.83.47:43781: write tcp 38.102.83.47:58720->38.102.83.47:43781: write: connection reset by peer Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.932463 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5rjsv"] Nov 24 19:50:04 crc kubenswrapper[5035]: E1124 19:50:04.933503 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="extract-content" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.933521 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="extract-content" Nov 24 19:50:04 crc kubenswrapper[5035]: E1124 19:50:04.933551 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="registry-server" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.933560 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="registry-server" Nov 24 19:50:04 crc kubenswrapper[5035]: E1124 19:50:04.933589 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="extract-utilities" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.933599 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="extract-utilities" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.933858 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbabe1be-acac-4c15-9465-df84149a2dde" containerName="registry-server" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.935793 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.945208 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rjsv"] Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.986806 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-utilities\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.987164 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-catalog-content\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:04 crc kubenswrapper[5035]: I1124 19:50:04.987203 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2pnn\" (UniqueName: \"kubernetes.io/projected/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-kube-api-access-x2pnn\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.088578 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-utilities\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.088715 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-catalog-content\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.088767 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2pnn\" (UniqueName: \"kubernetes.io/projected/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-kube-api-access-x2pnn\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.089558 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-utilities\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.089831 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-catalog-content\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.110987 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2pnn\" (UniqueName: \"kubernetes.io/projected/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-kube-api-access-x2pnn\") pod \"redhat-marketplace-5rjsv\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.258520 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:05 crc kubenswrapper[5035]: I1124 19:50:05.722542 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rjsv"] Nov 24 19:50:06 crc kubenswrapper[5035]: I1124 19:50:06.383924 5035 generic.go:334] "Generic (PLEG): container finished" podID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerID="75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0" exitCode=0 Nov 24 19:50:06 crc kubenswrapper[5035]: I1124 19:50:06.384026 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rjsv" event={"ID":"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90","Type":"ContainerDied","Data":"75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0"} Nov 24 19:50:06 crc kubenswrapper[5035]: I1124 19:50:06.384237 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rjsv" event={"ID":"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90","Type":"ContainerStarted","Data":"e6b1ab70bda42db6f27cdf5e66c48d8ddb1a8dc865513cd9c80de7810be5c5d7"} Nov 24 19:50:07 crc kubenswrapper[5035]: I1124 19:50:07.399269 5035 generic.go:334] "Generic (PLEG): container finished" podID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerID="5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641" exitCode=0 Nov 24 19:50:07 crc kubenswrapper[5035]: I1124 19:50:07.399798 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rjsv" event={"ID":"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90","Type":"ContainerDied","Data":"5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641"} Nov 24 19:50:08 crc kubenswrapper[5035]: I1124 19:50:08.408737 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rjsv" event={"ID":"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90","Type":"ContainerStarted","Data":"24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac"} Nov 24 19:50:08 crc kubenswrapper[5035]: I1124 19:50:08.434812 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5rjsv" podStartSLOduration=2.985071675 podStartE2EDuration="4.434796675s" podCreationTimestamp="2025-11-24 19:50:04 +0000 UTC" firstStartedPulling="2025-11-24 19:50:06.386008081 +0000 UTC m=+2144.908514368" lastFinishedPulling="2025-11-24 19:50:07.835733111 +0000 UTC m=+2146.358239368" observedRunningTime="2025-11-24 19:50:08.431171367 +0000 UTC m=+2146.953677624" watchObservedRunningTime="2025-11-24 19:50:08.434796675 +0000 UTC m=+2146.957302922" Nov 24 19:50:11 crc kubenswrapper[5035]: E1124 19:50:11.047844 5035 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.47:33770->38.102.83.47:43781: read tcp 38.102.83.47:33770->38.102.83.47:43781: read: connection reset by peer Nov 24 19:50:15 crc kubenswrapper[5035]: I1124 19:50:15.259682 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:15 crc kubenswrapper[5035]: I1124 19:50:15.260284 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:15 crc kubenswrapper[5035]: I1124 19:50:15.336740 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:15 crc kubenswrapper[5035]: I1124 19:50:15.516936 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:15 crc kubenswrapper[5035]: I1124 19:50:15.574630 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rjsv"] Nov 24 19:50:17 crc kubenswrapper[5035]: I1124 19:50:17.496678 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5rjsv" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="registry-server" containerID="cri-o://24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac" gracePeriod=2 Nov 24 19:50:17 crc kubenswrapper[5035]: I1124 19:50:17.953542 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.140856 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2pnn\" (UniqueName: \"kubernetes.io/projected/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-kube-api-access-x2pnn\") pod \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.140916 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-utilities\") pod \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.140978 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-catalog-content\") pod \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\" (UID: \"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90\") " Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.142437 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-utilities" (OuterVolumeSpecName: "utilities") pod "619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" (UID: "619a7195-cd72-4f8b-aa3b-72b4fd7c5d90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.147201 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-kube-api-access-x2pnn" (OuterVolumeSpecName: "kube-api-access-x2pnn") pod "619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" (UID: "619a7195-cd72-4f8b-aa3b-72b4fd7c5d90"). InnerVolumeSpecName "kube-api-access-x2pnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.167806 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" (UID: "619a7195-cd72-4f8b-aa3b-72b4fd7c5d90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.242329 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.242977 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.245531 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2pnn\" (UniqueName: \"kubernetes.io/projected/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-kube-api-access-x2pnn\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.245579 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.245589 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.248884 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbqlq"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.257754 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-d8dhn"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.266443 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.272771 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4qf8z"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.280727 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.287320 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.294201 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-vgfmw"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.301138 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.308615 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.314510 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.323200 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-28n87"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.330797 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4qf8z"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.337309 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9ngj7"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.343329 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2lzsx"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.349225 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.354954 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-7lqf8"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.361048 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9f772"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.367955 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-26d94"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.509343 5035 generic.go:334] "Generic (PLEG): container finished" podID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerID="24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac" exitCode=0 Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.509413 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rjsv" event={"ID":"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90","Type":"ContainerDied","Data":"24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac"} Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.509480 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rjsv" event={"ID":"619a7195-cd72-4f8b-aa3b-72b4fd7c5d90","Type":"ContainerDied","Data":"e6b1ab70bda42db6f27cdf5e66c48d8ddb1a8dc865513cd9c80de7810be5c5d7"} Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.509509 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rjsv" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.509512 5035 scope.go:117] "RemoveContainer" containerID="24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.545423 5035 scope.go:117] "RemoveContainer" containerID="5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.546431 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rjsv"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.557980 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rjsv"] Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.572801 5035 scope.go:117] "RemoveContainer" containerID="75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.605610 5035 scope.go:117] "RemoveContainer" containerID="24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac" Nov 24 19:50:18 crc kubenswrapper[5035]: E1124 19:50:18.606435 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac\": container with ID starting with 24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac not found: ID does not exist" containerID="24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.606465 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac"} err="failed to get container status \"24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac\": rpc error: code = NotFound desc = could not find container \"24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac\": container with ID starting with 24c382de9330f5feb9ae52e1342e6340e07c0d3b13d139033ea3b0344916abac not found: ID does not exist" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.606485 5035 scope.go:117] "RemoveContainer" containerID="5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641" Nov 24 19:50:18 crc kubenswrapper[5035]: E1124 19:50:18.606920 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641\": container with ID starting with 5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641 not found: ID does not exist" containerID="5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.606976 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641"} err="failed to get container status \"5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641\": rpc error: code = NotFound desc = could not find container \"5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641\": container with ID starting with 5ec26d217cf4cfd359f4dcd89d806ecf89d56e548743be36a2f91211a13ea641 not found: ID does not exist" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.607025 5035 scope.go:117] "RemoveContainer" containerID="75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0" Nov 24 19:50:18 crc kubenswrapper[5035]: E1124 19:50:18.607509 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0\": container with ID starting with 75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0 not found: ID does not exist" containerID="75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0" Nov 24 19:50:18 crc kubenswrapper[5035]: I1124 19:50:18.607605 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0"} err="failed to get container status \"75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0\": rpc error: code = NotFound desc = could not find container \"75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0\": container with ID starting with 75dd3c38acaa6fbf36b7692878b78c6f2b7040fb5aa3bd57145e6ce4166bc7d0 not found: ID does not exist" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.209732 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ee29809-634e-4c86-9605-5368bfc80b36" path="/var/lib/kubelet/pods/5ee29809-634e-4c86-9605-5368bfc80b36/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.210564 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" path="/var/lib/kubelet/pods/619a7195-cd72-4f8b-aa3b-72b4fd7c5d90/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.211286 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82a1649c-9b91-4ed5-bff3-acddd2292b4f" path="/var/lib/kubelet/pods/82a1649c-9b91-4ed5-bff3-acddd2292b4f/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.212600 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="878700e6-143e-4bb4-a300-13fd3ce143ac" path="/var/lib/kubelet/pods/878700e6-143e-4bb4-a300-13fd3ce143ac/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.213320 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="945f0c1b-86e7-4759-96fb-cc5278b03215" path="/var/lib/kubelet/pods/945f0c1b-86e7-4759-96fb-cc5278b03215/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.213865 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ef3873-f86b-41de-9a54-96e8c47e5e16" path="/var/lib/kubelet/pods/a7ef3873-f86b-41de-9a54-96e8c47e5e16/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.214474 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bda65671-049d-4645-96fc-f6e582e70fcc" path="/var/lib/kubelet/pods/bda65671-049d-4645-96fc-f6e582e70fcc/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.215477 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3eadfa2-fc9a-49af-8e88-1c83793f075a" path="/var/lib/kubelet/pods/d3eadfa2-fc9a-49af-8e88-1c83793f075a/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.216129 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d74ff699-d2f6-4adc-b70a-b1a35e2d997a" path="/var/lib/kubelet/pods/d74ff699-d2f6-4adc-b70a-b1a35e2d997a/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.216742 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb570109-faff-4cd5-97ac-b9d0ac5398fb" path="/var/lib/kubelet/pods/eb570109-faff-4cd5-97ac-b9d0ac5398fb/volumes" Nov 24 19:50:20 crc kubenswrapper[5035]: I1124 19:50:20.217913 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec6e4be0-9f98-41d1-8bba-563fe9b39c78" path="/var/lib/kubelet/pods/ec6e4be0-9f98-41d1-8bba-563fe9b39c78/volumes" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.329536 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg"] Nov 24 19:50:24 crc kubenswrapper[5035]: E1124 19:50:24.330423 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="extract-utilities" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.330438 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="extract-utilities" Nov 24 19:50:24 crc kubenswrapper[5035]: E1124 19:50:24.330452 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="registry-server" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.330459 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="registry-server" Nov 24 19:50:24 crc kubenswrapper[5035]: E1124 19:50:24.330491 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="extract-content" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.330498 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="extract-content" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.330709 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="619a7195-cd72-4f8b-aa3b-72b4fd7c5d90" containerName="registry-server" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.331404 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.334089 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.337000 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.337013 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.337125 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.338039 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.351259 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg"] Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.461117 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.461524 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.461626 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhmph\" (UniqueName: \"kubernetes.io/projected/c6da5217-1cd1-4f4e-8636-bb4a3604e939-kube-api-access-mhmph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.461783 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.461824 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.563400 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.563493 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.563569 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.563595 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.563640 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhmph\" (UniqueName: \"kubernetes.io/projected/c6da5217-1cd1-4f4e-8636-bb4a3604e939-kube-api-access-mhmph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.573162 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.574055 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.575609 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.577718 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.585249 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhmph\" (UniqueName: \"kubernetes.io/projected/c6da5217-1cd1-4f4e-8636-bb4a3604e939-kube-api-access-mhmph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:24 crc kubenswrapper[5035]: I1124 19:50:24.670159 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:25 crc kubenswrapper[5035]: I1124 19:50:25.210630 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg"] Nov 24 19:50:25 crc kubenswrapper[5035]: I1124 19:50:25.593332 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" event={"ID":"c6da5217-1cd1-4f4e-8636-bb4a3604e939","Type":"ContainerStarted","Data":"1d79d193ff5d97db459aaf378eb6965ff3f5a90a1bddd71fd828e117d8de0e4a"} Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.552347 5035 scope.go:117] "RemoveContainer" containerID="7a6b96b6aff7d81089fb7bc81d6f7699e44fd904b924bddda69e1d51a6f701b5" Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.603364 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" event={"ID":"c6da5217-1cd1-4f4e-8636-bb4a3604e939","Type":"ContainerStarted","Data":"76803e432f97189c3a7ed9d03b6a42fb311e897d3fbb9bb3b48da365efd77ca0"} Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.605100 5035 scope.go:117] "RemoveContainer" containerID="2d964ff0dc045abf4c4834bf3939982aea08ab94133bed49ec875a2f9ea9e82a" Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.623887 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" podStartSLOduration=1.885995904 podStartE2EDuration="2.623859572s" podCreationTimestamp="2025-11-24 19:50:24 +0000 UTC" firstStartedPulling="2025-11-24 19:50:25.214678111 +0000 UTC m=+2163.737184368" lastFinishedPulling="2025-11-24 19:50:25.952541739 +0000 UTC m=+2164.475048036" observedRunningTime="2025-11-24 19:50:26.615601968 +0000 UTC m=+2165.138108235" watchObservedRunningTime="2025-11-24 19:50:26.623859572 +0000 UTC m=+2165.146365869" Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.642500 5035 scope.go:117] "RemoveContainer" containerID="a28a1ad6b71c735874f3eec6177bffc3109050a8d7da8bd5407c3ee0381078d6" Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.671658 5035 scope.go:117] "RemoveContainer" containerID="3d451ec7a73ffb3c6ebc60f845290c24e9b8c584d9d3251c9733ea3586004d03" Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.729347 5035 scope.go:117] "RemoveContainer" containerID="400090af98e3ce470f10bb5d9bfcf877e2289c5688d7ca8eddfb4a3201f4bcaa" Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.786376 5035 scope.go:117] "RemoveContainer" containerID="1ae38fe2079d3cb2feb21ffbbe6deb6997d914223fdc20ece9b53910be21b0c3" Nov 24 19:50:26 crc kubenswrapper[5035]: I1124 19:50:26.829779 5035 scope.go:117] "RemoveContainer" containerID="7107bfbed7b0fc114c81497ac87bcb2ddecd6b064f06f2d6a4d76b0601f54215" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.441348 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zqr9x"] Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.449319 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.460217 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqr9x"] Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.568948 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrk6r\" (UniqueName: \"kubernetes.io/projected/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-kube-api-access-vrk6r\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.569054 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-catalog-content\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.569127 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-utilities\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.672380 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrk6r\" (UniqueName: \"kubernetes.io/projected/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-kube-api-access-vrk6r\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.672486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-catalog-content\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.672555 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-utilities\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.673078 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-utilities\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.673271 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-catalog-content\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.702904 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrk6r\" (UniqueName: \"kubernetes.io/projected/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-kube-api-access-vrk6r\") pod \"community-operators-zqr9x\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:39 crc kubenswrapper[5035]: I1124 19:50:39.797680 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:40 crc kubenswrapper[5035]: I1124 19:50:40.317076 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqr9x"] Nov 24 19:50:40 crc kubenswrapper[5035]: W1124 19:50:40.325577 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8224a9e2_e84a_4c43_adb4_6afd1b8414e7.slice/crio-a13dd83f85099a8fadc9545560e768e6d6bc32a3e8ad6147372237746948729e WatchSource:0}: Error finding container a13dd83f85099a8fadc9545560e768e6d6bc32a3e8ad6147372237746948729e: Status 404 returned error can't find the container with id a13dd83f85099a8fadc9545560e768e6d6bc32a3e8ad6147372237746948729e Nov 24 19:50:40 crc kubenswrapper[5035]: I1124 19:50:40.759939 5035 generic.go:334] "Generic (PLEG): container finished" podID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerID="583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d" exitCode=0 Nov 24 19:50:40 crc kubenswrapper[5035]: I1124 19:50:40.760181 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqr9x" event={"ID":"8224a9e2-e84a-4c43-adb4-6afd1b8414e7","Type":"ContainerDied","Data":"583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d"} Nov 24 19:50:40 crc kubenswrapper[5035]: I1124 19:50:40.760351 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqr9x" event={"ID":"8224a9e2-e84a-4c43-adb4-6afd1b8414e7","Type":"ContainerStarted","Data":"a13dd83f85099a8fadc9545560e768e6d6bc32a3e8ad6147372237746948729e"} Nov 24 19:50:42 crc kubenswrapper[5035]: I1124 19:50:42.790954 5035 generic.go:334] "Generic (PLEG): container finished" podID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerID="8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21" exitCode=0 Nov 24 19:50:42 crc kubenswrapper[5035]: I1124 19:50:42.791020 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqr9x" event={"ID":"8224a9e2-e84a-4c43-adb4-6afd1b8414e7","Type":"ContainerDied","Data":"8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21"} Nov 24 19:50:45 crc kubenswrapper[5035]: I1124 19:50:45.235014 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:50:45 crc kubenswrapper[5035]: I1124 19:50:45.235747 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:50:45 crc kubenswrapper[5035]: I1124 19:50:45.825645 5035 generic.go:334] "Generic (PLEG): container finished" podID="c6da5217-1cd1-4f4e-8636-bb4a3604e939" containerID="76803e432f97189c3a7ed9d03b6a42fb311e897d3fbb9bb3b48da365efd77ca0" exitCode=0 Nov 24 19:50:45 crc kubenswrapper[5035]: I1124 19:50:45.825720 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" event={"ID":"c6da5217-1cd1-4f4e-8636-bb4a3604e939","Type":"ContainerDied","Data":"76803e432f97189c3a7ed9d03b6a42fb311e897d3fbb9bb3b48da365efd77ca0"} Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.350755 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.535765 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhmph\" (UniqueName: \"kubernetes.io/projected/c6da5217-1cd1-4f4e-8636-bb4a3604e939-kube-api-access-mhmph\") pod \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.536122 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-inventory\") pod \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.536406 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ssh-key\") pod \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.536526 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ceph\") pod \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.536698 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-repo-setup-combined-ca-bundle\") pod \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\" (UID: \"c6da5217-1cd1-4f4e-8636-bb4a3604e939\") " Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.542067 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c6da5217-1cd1-4f4e-8636-bb4a3604e939" (UID: "c6da5217-1cd1-4f4e-8636-bb4a3604e939"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.543420 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ceph" (OuterVolumeSpecName: "ceph") pod "c6da5217-1cd1-4f4e-8636-bb4a3604e939" (UID: "c6da5217-1cd1-4f4e-8636-bb4a3604e939"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.544203 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6da5217-1cd1-4f4e-8636-bb4a3604e939-kube-api-access-mhmph" (OuterVolumeSpecName: "kube-api-access-mhmph") pod "c6da5217-1cd1-4f4e-8636-bb4a3604e939" (UID: "c6da5217-1cd1-4f4e-8636-bb4a3604e939"). InnerVolumeSpecName "kube-api-access-mhmph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.564858 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-inventory" (OuterVolumeSpecName: "inventory") pod "c6da5217-1cd1-4f4e-8636-bb4a3604e939" (UID: "c6da5217-1cd1-4f4e-8636-bb4a3604e939"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.575959 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6da5217-1cd1-4f4e-8636-bb4a3604e939" (UID: "c6da5217-1cd1-4f4e-8636-bb4a3604e939"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.638958 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhmph\" (UniqueName: \"kubernetes.io/projected/c6da5217-1cd1-4f4e-8636-bb4a3604e939-kube-api-access-mhmph\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.639000 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.639013 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.639025 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.639038 5035 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6da5217-1cd1-4f4e-8636-bb4a3604e939-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.849873 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqr9x" event={"ID":"8224a9e2-e84a-4c43-adb4-6afd1b8414e7","Type":"ContainerStarted","Data":"fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2"} Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.855376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" event={"ID":"c6da5217-1cd1-4f4e-8636-bb4a3604e939","Type":"ContainerDied","Data":"1d79d193ff5d97db459aaf378eb6965ff3f5a90a1bddd71fd828e117d8de0e4a"} Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.855416 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d79d193ff5d97db459aaf378eb6965ff3f5a90a1bddd71fd828e117d8de0e4a" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.855565 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.900809 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zqr9x" podStartSLOduration=2.440410632 podStartE2EDuration="8.90078679s" podCreationTimestamp="2025-11-24 19:50:39 +0000 UTC" firstStartedPulling="2025-11-24 19:50:40.762095089 +0000 UTC m=+2179.284601356" lastFinishedPulling="2025-11-24 19:50:47.222471257 +0000 UTC m=+2185.744977514" observedRunningTime="2025-11-24 19:50:47.879864743 +0000 UTC m=+2186.402371020" watchObservedRunningTime="2025-11-24 19:50:47.90078679 +0000 UTC m=+2186.423293067" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.956643 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9"] Nov 24 19:50:47 crc kubenswrapper[5035]: E1124 19:50:47.957055 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6da5217-1cd1-4f4e-8636-bb4a3604e939" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.957080 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6da5217-1cd1-4f4e-8636-bb4a3604e939" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.957357 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6da5217-1cd1-4f4e-8636-bb4a3604e939" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.958051 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.963078 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.963711 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.964184 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.964450 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.964748 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:50:47 crc kubenswrapper[5035]: I1124 19:50:47.967219 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9"] Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.147770 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.147866 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.147889 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.147908 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.147929 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97tr4\" (UniqueName: \"kubernetes.io/projected/8f7443fa-cea8-4010-8cac-0186257581cb-kube-api-access-97tr4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.250541 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.250724 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.250761 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.250807 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.250846 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97tr4\" (UniqueName: \"kubernetes.io/projected/8f7443fa-cea8-4010-8cac-0186257581cb-kube-api-access-97tr4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.258893 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.258980 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.262982 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.263092 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.272252 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97tr4\" (UniqueName: \"kubernetes.io/projected/8f7443fa-cea8-4010-8cac-0186257581cb-kube-api-access-97tr4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.321114 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:50:48 crc kubenswrapper[5035]: I1124 19:50:48.867080 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9"] Nov 24 19:50:49 crc kubenswrapper[5035]: I1124 19:50:49.798389 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:49 crc kubenswrapper[5035]: I1124 19:50:49.799033 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:49 crc kubenswrapper[5035]: I1124 19:50:49.855008 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:49 crc kubenswrapper[5035]: I1124 19:50:49.893442 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" event={"ID":"8f7443fa-cea8-4010-8cac-0186257581cb","Type":"ContainerStarted","Data":"0faef72c260cec97ad8f758b4efb5e335c828463f55f598db326c398a8b45a52"} Nov 24 19:50:49 crc kubenswrapper[5035]: I1124 19:50:49.893517 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" event={"ID":"8f7443fa-cea8-4010-8cac-0186257581cb","Type":"ContainerStarted","Data":"5b411ec6dd341afbf56a84ce431d113211ef1a639165db21199b3f3678c064e6"} Nov 24 19:50:49 crc kubenswrapper[5035]: I1124 19:50:49.909794 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" podStartSLOduration=2.462183689 podStartE2EDuration="2.909776035s" podCreationTimestamp="2025-11-24 19:50:47 +0000 UTC" firstStartedPulling="2025-11-24 19:50:48.867795962 +0000 UTC m=+2187.390302219" lastFinishedPulling="2025-11-24 19:50:49.315388308 +0000 UTC m=+2187.837894565" observedRunningTime="2025-11-24 19:50:49.906489426 +0000 UTC m=+2188.428995683" watchObservedRunningTime="2025-11-24 19:50:49.909776035 +0000 UTC m=+2188.432282292" Nov 24 19:50:59 crc kubenswrapper[5035]: I1124 19:50:59.856544 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:50:59 crc kubenswrapper[5035]: I1124 19:50:59.909366 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqr9x"] Nov 24 19:50:59 crc kubenswrapper[5035]: I1124 19:50:59.988787 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zqr9x" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="registry-server" containerID="cri-o://fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2" gracePeriod=2 Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.486794 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.617114 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrk6r\" (UniqueName: \"kubernetes.io/projected/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-kube-api-access-vrk6r\") pod \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.617237 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-catalog-content\") pod \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.617396 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-utilities\") pod \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\" (UID: \"8224a9e2-e84a-4c43-adb4-6afd1b8414e7\") " Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.618188 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-utilities" (OuterVolumeSpecName: "utilities") pod "8224a9e2-e84a-4c43-adb4-6afd1b8414e7" (UID: "8224a9e2-e84a-4c43-adb4-6afd1b8414e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.622659 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-kube-api-access-vrk6r" (OuterVolumeSpecName: "kube-api-access-vrk6r") pod "8224a9e2-e84a-4c43-adb4-6afd1b8414e7" (UID: "8224a9e2-e84a-4c43-adb4-6afd1b8414e7"). InnerVolumeSpecName "kube-api-access-vrk6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.684747 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8224a9e2-e84a-4c43-adb4-6afd1b8414e7" (UID: "8224a9e2-e84a-4c43-adb4-6afd1b8414e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.719614 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrk6r\" (UniqueName: \"kubernetes.io/projected/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-kube-api-access-vrk6r\") on node \"crc\" DevicePath \"\"" Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.719647 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:51:00 crc kubenswrapper[5035]: I1124 19:51:00.719661 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8224a9e2-e84a-4c43-adb4-6afd1b8414e7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.000738 5035 generic.go:334] "Generic (PLEG): container finished" podID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerID="fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2" exitCode=0 Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.000818 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqr9x" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.000815 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqr9x" event={"ID":"8224a9e2-e84a-4c43-adb4-6afd1b8414e7","Type":"ContainerDied","Data":"fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2"} Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.001420 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqr9x" event={"ID":"8224a9e2-e84a-4c43-adb4-6afd1b8414e7","Type":"ContainerDied","Data":"a13dd83f85099a8fadc9545560e768e6d6bc32a3e8ad6147372237746948729e"} Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.001458 5035 scope.go:117] "RemoveContainer" containerID="fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.024889 5035 scope.go:117] "RemoveContainer" containerID="8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.044184 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqr9x"] Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.049129 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zqr9x"] Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.056090 5035 scope.go:117] "RemoveContainer" containerID="583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.104925 5035 scope.go:117] "RemoveContainer" containerID="fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2" Nov 24 19:51:01 crc kubenswrapper[5035]: E1124 19:51:01.105395 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2\": container with ID starting with fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2 not found: ID does not exist" containerID="fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.105424 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2"} err="failed to get container status \"fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2\": rpc error: code = NotFound desc = could not find container \"fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2\": container with ID starting with fc7cf8589a0cdf547ba406e198afe0fd98213462de8a80f93bad82ca718f16e2 not found: ID does not exist" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.105444 5035 scope.go:117] "RemoveContainer" containerID="8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21" Nov 24 19:51:01 crc kubenswrapper[5035]: E1124 19:51:01.105733 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21\": container with ID starting with 8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21 not found: ID does not exist" containerID="8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.105749 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21"} err="failed to get container status \"8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21\": rpc error: code = NotFound desc = could not find container \"8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21\": container with ID starting with 8a5575af9f8f20aefba67dd8dcda70c8b72f3b4a93b72569add03285ab634d21 not found: ID does not exist" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.105761 5035 scope.go:117] "RemoveContainer" containerID="583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d" Nov 24 19:51:01 crc kubenswrapper[5035]: E1124 19:51:01.106134 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d\": container with ID starting with 583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d not found: ID does not exist" containerID="583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d" Nov 24 19:51:01 crc kubenswrapper[5035]: I1124 19:51:01.106174 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d"} err="failed to get container status \"583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d\": rpc error: code = NotFound desc = could not find container \"583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d\": container with ID starting with 583bb0b852fdbc9e317a0810fe6fee46fd6b31a51c3d59934f18e4343801303d not found: ID does not exist" Nov 24 19:51:02 crc kubenswrapper[5035]: I1124 19:51:02.223505 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" path="/var/lib/kubelet/pods/8224a9e2-e84a-4c43-adb4-6afd1b8414e7/volumes" Nov 24 19:51:15 crc kubenswrapper[5035]: I1124 19:51:15.234180 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:51:15 crc kubenswrapper[5035]: I1124 19:51:15.234975 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:51:27 crc kubenswrapper[5035]: I1124 19:51:27.016914 5035 scope.go:117] "RemoveContainer" containerID="45a09b5f169447bad1fd6d005942767fc7cefb67e285cf436876df57268d0278" Nov 24 19:51:27 crc kubenswrapper[5035]: I1124 19:51:27.079223 5035 scope.go:117] "RemoveContainer" containerID="20ea6af7e5817c8b7c8cfa8ed5ce976d5e502040602f74bbf67cfd06056ed2f1" Nov 24 19:51:27 crc kubenswrapper[5035]: I1124 19:51:27.132410 5035 scope.go:117] "RemoveContainer" containerID="3f336398724b57b54db8dcd327d90f55c7184faec5ac1bd3368732acda2caca4" Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.234747 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.235378 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.235436 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.236340 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.236402 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" gracePeriod=600 Nov 24 19:51:45 crc kubenswrapper[5035]: E1124 19:51:45.390066 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.444503 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" exitCode=0 Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.444558 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20"} Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.444618 5035 scope.go:117] "RemoveContainer" containerID="614af86f0ff3c54135bc9d250592e150a845656f46635f75ac952947e7986d0f" Nov 24 19:51:45 crc kubenswrapper[5035]: I1124 19:51:45.445241 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:51:45 crc kubenswrapper[5035]: E1124 19:51:45.447703 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:51:56 crc kubenswrapper[5035]: I1124 19:51:56.200442 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:51:56 crc kubenswrapper[5035]: E1124 19:51:56.201245 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:52:10 crc kubenswrapper[5035]: I1124 19:52:10.201446 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:52:10 crc kubenswrapper[5035]: E1124 19:52:10.202697 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:52:21 crc kubenswrapper[5035]: I1124 19:52:21.200231 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:52:21 crc kubenswrapper[5035]: E1124 19:52:21.201464 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:52:34 crc kubenswrapper[5035]: I1124 19:52:34.201354 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:52:34 crc kubenswrapper[5035]: E1124 19:52:34.202423 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:52:45 crc kubenswrapper[5035]: I1124 19:52:45.200452 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:52:45 crc kubenswrapper[5035]: E1124 19:52:45.202939 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:53:00 crc kubenswrapper[5035]: I1124 19:53:00.203725 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:53:00 crc kubenswrapper[5035]: E1124 19:53:00.204916 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:53:11 crc kubenswrapper[5035]: I1124 19:53:11.200726 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:53:11 crc kubenswrapper[5035]: E1124 19:53:11.201989 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:53:24 crc kubenswrapper[5035]: I1124 19:53:24.200625 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:53:24 crc kubenswrapper[5035]: E1124 19:53:24.201442 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:53:38 crc kubenswrapper[5035]: I1124 19:53:38.200065 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:53:38 crc kubenswrapper[5035]: E1124 19:53:38.200897 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:53:53 crc kubenswrapper[5035]: I1124 19:53:53.202461 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:53:53 crc kubenswrapper[5035]: E1124 19:53:53.203454 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:54:05 crc kubenswrapper[5035]: I1124 19:54:05.201963 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:54:05 crc kubenswrapper[5035]: E1124 19:54:05.203018 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:54:19 crc kubenswrapper[5035]: I1124 19:54:19.201877 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:54:19 crc kubenswrapper[5035]: E1124 19:54:19.202866 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:54:32 crc kubenswrapper[5035]: I1124 19:54:32.215521 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:54:32 crc kubenswrapper[5035]: E1124 19:54:32.216452 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:54:44 crc kubenswrapper[5035]: I1124 19:54:44.200386 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:54:44 crc kubenswrapper[5035]: E1124 19:54:44.201375 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:54:59 crc kubenswrapper[5035]: I1124 19:54:59.200211 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:54:59 crc kubenswrapper[5035]: E1124 19:54:59.201100 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:55:12 crc kubenswrapper[5035]: I1124 19:55:12.207507 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:55:12 crc kubenswrapper[5035]: E1124 19:55:12.208862 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:55:20 crc kubenswrapper[5035]: I1124 19:55:20.529197 5035 generic.go:334] "Generic (PLEG): container finished" podID="8f7443fa-cea8-4010-8cac-0186257581cb" containerID="0faef72c260cec97ad8f758b4efb5e335c828463f55f598db326c398a8b45a52" exitCode=0 Nov 24 19:55:20 crc kubenswrapper[5035]: I1124 19:55:20.529325 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" event={"ID":"8f7443fa-cea8-4010-8cac-0186257581cb","Type":"ContainerDied","Data":"0faef72c260cec97ad8f758b4efb5e335c828463f55f598db326c398a8b45a52"} Nov 24 19:55:21 crc kubenswrapper[5035]: I1124 19:55:21.964861 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.059492 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-inventory\") pod \"8f7443fa-cea8-4010-8cac-0186257581cb\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.059530 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97tr4\" (UniqueName: \"kubernetes.io/projected/8f7443fa-cea8-4010-8cac-0186257581cb-kube-api-access-97tr4\") pod \"8f7443fa-cea8-4010-8cac-0186257581cb\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.059615 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ssh-key\") pod \"8f7443fa-cea8-4010-8cac-0186257581cb\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.059710 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ceph\") pod \"8f7443fa-cea8-4010-8cac-0186257581cb\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.059799 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-bootstrap-combined-ca-bundle\") pod \"8f7443fa-cea8-4010-8cac-0186257581cb\" (UID: \"8f7443fa-cea8-4010-8cac-0186257581cb\") " Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.065934 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8f7443fa-cea8-4010-8cac-0186257581cb" (UID: "8f7443fa-cea8-4010-8cac-0186257581cb"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.066039 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ceph" (OuterVolumeSpecName: "ceph") pod "8f7443fa-cea8-4010-8cac-0186257581cb" (UID: "8f7443fa-cea8-4010-8cac-0186257581cb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.066994 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f7443fa-cea8-4010-8cac-0186257581cb-kube-api-access-97tr4" (OuterVolumeSpecName: "kube-api-access-97tr4") pod "8f7443fa-cea8-4010-8cac-0186257581cb" (UID: "8f7443fa-cea8-4010-8cac-0186257581cb"). InnerVolumeSpecName "kube-api-access-97tr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.087819 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-inventory" (OuterVolumeSpecName: "inventory") pod "8f7443fa-cea8-4010-8cac-0186257581cb" (UID: "8f7443fa-cea8-4010-8cac-0186257581cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.094198 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8f7443fa-cea8-4010-8cac-0186257581cb" (UID: "8f7443fa-cea8-4010-8cac-0186257581cb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.162601 5035 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.162675 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.162700 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97tr4\" (UniqueName: \"kubernetes.io/projected/8f7443fa-cea8-4010-8cac-0186257581cb-kube-api-access-97tr4\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.162723 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.162748 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f7443fa-cea8-4010-8cac-0186257581cb-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.552313 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" event={"ID":"8f7443fa-cea8-4010-8cac-0186257581cb","Type":"ContainerDied","Data":"5b411ec6dd341afbf56a84ce431d113211ef1a639165db21199b3f3678c064e6"} Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.552516 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b411ec6dd341afbf56a84ce431d113211ef1a639165db21199b3f3678c064e6" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.552548 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.678132 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7"] Nov 24 19:55:22 crc kubenswrapper[5035]: E1124 19:55:22.678782 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="registry-server" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.678891 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="registry-server" Nov 24 19:55:22 crc kubenswrapper[5035]: E1124 19:55:22.678963 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="extract-utilities" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.679015 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="extract-utilities" Nov 24 19:55:22 crc kubenswrapper[5035]: E1124 19:55:22.679070 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f7443fa-cea8-4010-8cac-0186257581cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.679135 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f7443fa-cea8-4010-8cac-0186257581cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 19:55:22 crc kubenswrapper[5035]: E1124 19:55:22.679197 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="extract-content" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.679266 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="extract-content" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.679528 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f7443fa-cea8-4010-8cac-0186257581cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.679610 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8224a9e2-e84a-4c43-adb4-6afd1b8414e7" containerName="registry-server" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.680344 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.685554 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.686097 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7"] Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.685901 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.685556 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.685826 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.685884 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.774370 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.774415 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.774439 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhxd9\" (UniqueName: \"kubernetes.io/projected/8eb8573f-9078-4348-a9e3-d4156a32a917-kube-api-access-hhxd9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.774475 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.876229 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.876272 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.876313 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhxd9\" (UniqueName: \"kubernetes.io/projected/8eb8573f-9078-4348-a9e3-d4156a32a917-kube-api-access-hhxd9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.876350 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.881898 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.881960 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.883423 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:22 crc kubenswrapper[5035]: I1124 19:55:22.909589 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhxd9\" (UniqueName: \"kubernetes.io/projected/8eb8573f-9078-4348-a9e3-d4156a32a917-kube-api-access-hhxd9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:23 crc kubenswrapper[5035]: I1124 19:55:23.008672 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:23 crc kubenswrapper[5035]: I1124 19:55:23.657261 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 19:55:23 crc kubenswrapper[5035]: I1124 19:55:23.658416 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7"] Nov 24 19:55:24 crc kubenswrapper[5035]: I1124 19:55:24.572054 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" event={"ID":"8eb8573f-9078-4348-a9e3-d4156a32a917","Type":"ContainerStarted","Data":"781c3d889cbfe965dfaba84900ba118a0d7cfcc452decce56696ebf06a6917c9"} Nov 24 19:55:24 crc kubenswrapper[5035]: I1124 19:55:24.572592 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" event={"ID":"8eb8573f-9078-4348-a9e3-d4156a32a917","Type":"ContainerStarted","Data":"f127cf4ab45e30cceb09e461758a7e89940f7fbcc18aeb263d97277e3afb223d"} Nov 24 19:55:24 crc kubenswrapper[5035]: I1124 19:55:24.599958 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" podStartSLOduration=2.176224792 podStartE2EDuration="2.599933518s" podCreationTimestamp="2025-11-24 19:55:22 +0000 UTC" firstStartedPulling="2025-11-24 19:55:23.656911564 +0000 UTC m=+2462.179417831" lastFinishedPulling="2025-11-24 19:55:24.08062029 +0000 UTC m=+2462.603126557" observedRunningTime="2025-11-24 19:55:24.594655286 +0000 UTC m=+2463.117161583" watchObservedRunningTime="2025-11-24 19:55:24.599933518 +0000 UTC m=+2463.122439785" Nov 24 19:55:27 crc kubenswrapper[5035]: I1124 19:55:27.201090 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:55:27 crc kubenswrapper[5035]: E1124 19:55:27.201893 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:55:31 crc kubenswrapper[5035]: E1124 19:55:31.527083 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f7443fa_cea8_4010_8cac_0186257581cb.slice\": RecentStats: unable to find data in memory cache]" Nov 24 19:55:39 crc kubenswrapper[5035]: I1124 19:55:39.200720 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:55:39 crc kubenswrapper[5035]: E1124 19:55:39.201670 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:55:41 crc kubenswrapper[5035]: E1124 19:55:41.734714 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f7443fa_cea8_4010_8cac_0186257581cb.slice\": RecentStats: unable to find data in memory cache]" Nov 24 19:55:51 crc kubenswrapper[5035]: E1124 19:55:51.952219 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f7443fa_cea8_4010_8cac_0186257581cb.slice\": RecentStats: unable to find data in memory cache]" Nov 24 19:55:52 crc kubenswrapper[5035]: I1124 19:55:52.855564 5035 generic.go:334] "Generic (PLEG): container finished" podID="8eb8573f-9078-4348-a9e3-d4156a32a917" containerID="781c3d889cbfe965dfaba84900ba118a0d7cfcc452decce56696ebf06a6917c9" exitCode=0 Nov 24 19:55:52 crc kubenswrapper[5035]: I1124 19:55:52.855609 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" event={"ID":"8eb8573f-9078-4348-a9e3-d4156a32a917","Type":"ContainerDied","Data":"781c3d889cbfe965dfaba84900ba118a0d7cfcc452decce56696ebf06a6917c9"} Nov 24 19:55:53 crc kubenswrapper[5035]: I1124 19:55:53.200210 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:55:53 crc kubenswrapper[5035]: E1124 19:55:53.200509 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.352266 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.464572 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ceph\") pod \"8eb8573f-9078-4348-a9e3-d4156a32a917\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.464618 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ssh-key\") pod \"8eb8573f-9078-4348-a9e3-d4156a32a917\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.464722 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-inventory\") pod \"8eb8573f-9078-4348-a9e3-d4156a32a917\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.464753 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhxd9\" (UniqueName: \"kubernetes.io/projected/8eb8573f-9078-4348-a9e3-d4156a32a917-kube-api-access-hhxd9\") pod \"8eb8573f-9078-4348-a9e3-d4156a32a917\" (UID: \"8eb8573f-9078-4348-a9e3-d4156a32a917\") " Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.471911 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ceph" (OuterVolumeSpecName: "ceph") pod "8eb8573f-9078-4348-a9e3-d4156a32a917" (UID: "8eb8573f-9078-4348-a9e3-d4156a32a917"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.471984 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eb8573f-9078-4348-a9e3-d4156a32a917-kube-api-access-hhxd9" (OuterVolumeSpecName: "kube-api-access-hhxd9") pod "8eb8573f-9078-4348-a9e3-d4156a32a917" (UID: "8eb8573f-9078-4348-a9e3-d4156a32a917"). InnerVolumeSpecName "kube-api-access-hhxd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.492846 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8eb8573f-9078-4348-a9e3-d4156a32a917" (UID: "8eb8573f-9078-4348-a9e3-d4156a32a917"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.504435 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-inventory" (OuterVolumeSpecName: "inventory") pod "8eb8573f-9078-4348-a9e3-d4156a32a917" (UID: "8eb8573f-9078-4348-a9e3-d4156a32a917"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.566384 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.566457 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhxd9\" (UniqueName: \"kubernetes.io/projected/8eb8573f-9078-4348-a9e3-d4156a32a917-kube-api-access-hhxd9\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.566474 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.566519 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8eb8573f-9078-4348-a9e3-d4156a32a917-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.886223 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" event={"ID":"8eb8573f-9078-4348-a9e3-d4156a32a917","Type":"ContainerDied","Data":"f127cf4ab45e30cceb09e461758a7e89940f7fbcc18aeb263d97277e3afb223d"} Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.886717 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.886733 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f127cf4ab45e30cceb09e461758a7e89940f7fbcc18aeb263d97277e3afb223d" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.948079 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw"] Nov 24 19:55:54 crc kubenswrapper[5035]: E1124 19:55:54.948438 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb8573f-9078-4348-a9e3-d4156a32a917" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.948457 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb8573f-9078-4348-a9e3-d4156a32a917" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.948672 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eb8573f-9078-4348-a9e3-d4156a32a917" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.949354 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.951699 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.952033 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.952150 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.952397 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.954012 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.963680 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw"] Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.975323 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.975426 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.975468 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g896p\" (UniqueName: \"kubernetes.io/projected/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-kube-api-access-g896p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:54 crc kubenswrapper[5035]: I1124 19:55:54.975603 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.077783 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.077853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g896p\" (UniqueName: \"kubernetes.io/projected/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-kube-api-access-g896p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.077923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.077979 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.082405 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.082501 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.082536 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.094691 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g896p\" (UniqueName: \"kubernetes.io/projected/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-kube-api-access-g896p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.275834 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.829081 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw"] Nov 24 19:55:55 crc kubenswrapper[5035]: W1124 19:55:55.850837 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20cbe453_ecde_4214_835e_a2a6a1fb2cc3.slice/crio-2e10db1b5f33a6c4a7ab579bca8b9c4587e0b6f863b7d2f6e75a95d794775131 WatchSource:0}: Error finding container 2e10db1b5f33a6c4a7ab579bca8b9c4587e0b6f863b7d2f6e75a95d794775131: Status 404 returned error can't find the container with id 2e10db1b5f33a6c4a7ab579bca8b9c4587e0b6f863b7d2f6e75a95d794775131 Nov 24 19:55:55 crc kubenswrapper[5035]: I1124 19:55:55.899446 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" event={"ID":"20cbe453-ecde-4214-835e-a2a6a1fb2cc3","Type":"ContainerStarted","Data":"2e10db1b5f33a6c4a7ab579bca8b9c4587e0b6f863b7d2f6e75a95d794775131"} Nov 24 19:55:56 crc kubenswrapper[5035]: I1124 19:55:56.914164 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" event={"ID":"20cbe453-ecde-4214-835e-a2a6a1fb2cc3","Type":"ContainerStarted","Data":"4891e58c87e1e357e9aebffa2202e01bb2890e1f446427df6ad805e1eaf4eb32"} Nov 24 19:56:02 crc kubenswrapper[5035]: E1124 19:56:02.167763 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f7443fa_cea8_4010_8cac_0186257581cb.slice\": RecentStats: unable to find data in memory cache]" Nov 24 19:56:02 crc kubenswrapper[5035]: I1124 19:56:02.978518 5035 generic.go:334] "Generic (PLEG): container finished" podID="20cbe453-ecde-4214-835e-a2a6a1fb2cc3" containerID="4891e58c87e1e357e9aebffa2202e01bb2890e1f446427df6ad805e1eaf4eb32" exitCode=0 Nov 24 19:56:02 crc kubenswrapper[5035]: I1124 19:56:02.978835 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" event={"ID":"20cbe453-ecde-4214-835e-a2a6a1fb2cc3","Type":"ContainerDied","Data":"4891e58c87e1e357e9aebffa2202e01bb2890e1f446427df6ad805e1eaf4eb32"} Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.446431 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.470486 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g896p\" (UniqueName: \"kubernetes.io/projected/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-kube-api-access-g896p\") pod \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.470661 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ceph\") pod \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.470740 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-inventory\") pod \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.470776 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ssh-key\") pod \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\" (UID: \"20cbe453-ecde-4214-835e-a2a6a1fb2cc3\") " Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.506104 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ceph" (OuterVolumeSpecName: "ceph") pod "20cbe453-ecde-4214-835e-a2a6a1fb2cc3" (UID: "20cbe453-ecde-4214-835e-a2a6a1fb2cc3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.506316 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-kube-api-access-g896p" (OuterVolumeSpecName: "kube-api-access-g896p") pod "20cbe453-ecde-4214-835e-a2a6a1fb2cc3" (UID: "20cbe453-ecde-4214-835e-a2a6a1fb2cc3"). InnerVolumeSpecName "kube-api-access-g896p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.516763 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "20cbe453-ecde-4214-835e-a2a6a1fb2cc3" (UID: "20cbe453-ecde-4214-835e-a2a6a1fb2cc3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.520655 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-inventory" (OuterVolumeSpecName: "inventory") pod "20cbe453-ecde-4214-835e-a2a6a1fb2cc3" (UID: "20cbe453-ecde-4214-835e-a2a6a1fb2cc3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.573183 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g896p\" (UniqueName: \"kubernetes.io/projected/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-kube-api-access-g896p\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.573241 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.573259 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.573275 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20cbe453-ecde-4214-835e-a2a6a1fb2cc3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.995182 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" event={"ID":"20cbe453-ecde-4214-835e-a2a6a1fb2cc3","Type":"ContainerDied","Data":"2e10db1b5f33a6c4a7ab579bca8b9c4587e0b6f863b7d2f6e75a95d794775131"} Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.995216 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e10db1b5f33a6c4a7ab579bca8b9c4587e0b6f863b7d2f6e75a95d794775131" Nov 24 19:56:04 crc kubenswrapper[5035]: I1124 19:56:04.995246 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.068006 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r"] Nov 24 19:56:05 crc kubenswrapper[5035]: E1124 19:56:05.068627 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20cbe453-ecde-4214-835e-a2a6a1fb2cc3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.068660 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="20cbe453-ecde-4214-835e-a2a6a1fb2cc3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.069047 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="20cbe453-ecde-4214-835e-a2a6a1fb2cc3" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.070039 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.073100 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.073394 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.073467 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.073544 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.076213 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.084388 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.084449 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.084666 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.084770 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6pbq\" (UniqueName: \"kubernetes.io/projected/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-kube-api-access-r6pbq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.094071 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r"] Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.186327 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.186671 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6pbq\" (UniqueName: \"kubernetes.io/projected/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-kube-api-access-r6pbq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.187099 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.187264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.191819 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.192391 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.193342 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.203918 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6pbq\" (UniqueName: \"kubernetes.io/projected/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-kube-api-access-r6pbq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-klf6r\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.395581 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:05 crc kubenswrapper[5035]: I1124 19:56:05.938476 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r"] Nov 24 19:56:05 crc kubenswrapper[5035]: W1124 19:56:05.942967 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6f5bc3a_66aa_4f6a_b098_7b6a2512f120.slice/crio-eabc2001571e9b14dba3e00c8a1fc669ca7fc78a9339d4294bc1e9e370397cff WatchSource:0}: Error finding container eabc2001571e9b14dba3e00c8a1fc669ca7fc78a9339d4294bc1e9e370397cff: Status 404 returned error can't find the container with id eabc2001571e9b14dba3e00c8a1fc669ca7fc78a9339d4294bc1e9e370397cff Nov 24 19:56:06 crc kubenswrapper[5035]: I1124 19:56:06.003930 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" event={"ID":"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120","Type":"ContainerStarted","Data":"eabc2001571e9b14dba3e00c8a1fc669ca7fc78a9339d4294bc1e9e370397cff"} Nov 24 19:56:07 crc kubenswrapper[5035]: I1124 19:56:07.035869 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" event={"ID":"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120","Type":"ContainerStarted","Data":"e75a7a8dd1f2e08f0da24c2764a16e34b7cde5f67495c06177e792f70d805139"} Nov 24 19:56:07 crc kubenswrapper[5035]: I1124 19:56:07.065826 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" podStartSLOduration=1.639874305 podStartE2EDuration="2.065806461s" podCreationTimestamp="2025-11-24 19:56:05 +0000 UTC" firstStartedPulling="2025-11-24 19:56:05.945251843 +0000 UTC m=+2504.467758100" lastFinishedPulling="2025-11-24 19:56:06.371183999 +0000 UTC m=+2504.893690256" observedRunningTime="2025-11-24 19:56:07.056585684 +0000 UTC m=+2505.579091941" watchObservedRunningTime="2025-11-24 19:56:07.065806461 +0000 UTC m=+2505.588312718" Nov 24 19:56:08 crc kubenswrapper[5035]: I1124 19:56:08.200606 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:56:08 crc kubenswrapper[5035]: E1124 19:56:08.202402 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:56:12 crc kubenswrapper[5035]: E1124 19:56:12.426696 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f7443fa_cea8_4010_8cac_0186257581cb.slice\": RecentStats: unable to find data in memory cache]" Nov 24 19:56:23 crc kubenswrapper[5035]: I1124 19:56:23.199759 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:56:23 crc kubenswrapper[5035]: E1124 19:56:23.200503 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:56:37 crc kubenswrapper[5035]: I1124 19:56:37.200152 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:56:37 crc kubenswrapper[5035]: E1124 19:56:37.202117 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 19:56:48 crc kubenswrapper[5035]: I1124 19:56:48.200834 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 19:56:49 crc kubenswrapper[5035]: I1124 19:56:49.434985 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"25c3576c8f6c0c703275007665c285ca572de78d4e00d849bb0e88dffe634030"} Nov 24 19:56:50 crc kubenswrapper[5035]: I1124 19:56:50.446970 5035 generic.go:334] "Generic (PLEG): container finished" podID="e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" containerID="e75a7a8dd1f2e08f0da24c2764a16e34b7cde5f67495c06177e792f70d805139" exitCode=0 Nov 24 19:56:50 crc kubenswrapper[5035]: I1124 19:56:50.447089 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" event={"ID":"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120","Type":"ContainerDied","Data":"e75a7a8dd1f2e08f0da24c2764a16e34b7cde5f67495c06177e792f70d805139"} Nov 24 19:56:51 crc kubenswrapper[5035]: I1124 19:56:51.892541 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.058517 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6pbq\" (UniqueName: \"kubernetes.io/projected/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-kube-api-access-r6pbq\") pod \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.058719 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ssh-key\") pod \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.058813 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-inventory\") pod \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.058899 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ceph\") pod \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\" (UID: \"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120\") " Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.066104 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-kube-api-access-r6pbq" (OuterVolumeSpecName: "kube-api-access-r6pbq") pod "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" (UID: "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120"). InnerVolumeSpecName "kube-api-access-r6pbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.066908 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ceph" (OuterVolumeSpecName: "ceph") pod "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" (UID: "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.103510 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-inventory" (OuterVolumeSpecName: "inventory") pod "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" (UID: "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.106019 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" (UID: "e6f5bc3a-66aa-4f6a-b098-7b6a2512f120"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.160484 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.160515 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.160524 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.160532 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6pbq\" (UniqueName: \"kubernetes.io/projected/e6f5bc3a-66aa-4f6a-b098-7b6a2512f120-kube-api-access-r6pbq\") on node \"crc\" DevicePath \"\"" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.462834 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" event={"ID":"e6f5bc3a-66aa-4f6a-b098-7b6a2512f120","Type":"ContainerDied","Data":"eabc2001571e9b14dba3e00c8a1fc669ca7fc78a9339d4294bc1e9e370397cff"} Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.462872 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eabc2001571e9b14dba3e00c8a1fc669ca7fc78a9339d4294bc1e9e370397cff" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.462914 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-klf6r" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.583046 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x"] Nov 24 19:56:52 crc kubenswrapper[5035]: E1124 19:56:52.583707 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.583728 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.583918 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6f5bc3a-66aa-4f6a-b098-7b6a2512f120" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.584631 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.587878 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.588034 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.588154 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.589845 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.590849 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.592615 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x"] Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.668493 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.668591 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.668740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfmgn\" (UniqueName: \"kubernetes.io/projected/6ef15736-c842-4a20-b05f-9105eef653a7-kube-api-access-xfmgn\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.668790 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.770688 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfmgn\" (UniqueName: \"kubernetes.io/projected/6ef15736-c842-4a20-b05f-9105eef653a7-kube-api-access-xfmgn\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.770785 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.770929 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.771048 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.776643 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.776817 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.779149 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.790606 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfmgn\" (UniqueName: \"kubernetes.io/projected/6ef15736-c842-4a20-b05f-9105eef653a7-kube-api-access-xfmgn\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:52 crc kubenswrapper[5035]: I1124 19:56:52.943322 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:56:53 crc kubenswrapper[5035]: I1124 19:56:53.497108 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x"] Nov 24 19:56:54 crc kubenswrapper[5035]: I1124 19:56:54.480950 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" event={"ID":"6ef15736-c842-4a20-b05f-9105eef653a7","Type":"ContainerStarted","Data":"315d04c48f9fdbbea5c8dd5a0108decabe94e027cb666cb2ff91de3bbfbd18b0"} Nov 24 19:56:54 crc kubenswrapper[5035]: I1124 19:56:54.481593 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" event={"ID":"6ef15736-c842-4a20-b05f-9105eef653a7","Type":"ContainerStarted","Data":"29db042fe359fbfcdeed0fea14a02cb45548a71691b6533d8113ff67c8011e87"} Nov 24 19:56:54 crc kubenswrapper[5035]: I1124 19:56:54.504392 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" podStartSLOduration=2.066813955 podStartE2EDuration="2.504368417s" podCreationTimestamp="2025-11-24 19:56:52 +0000 UTC" firstStartedPulling="2025-11-24 19:56:53.501932572 +0000 UTC m=+2552.024438829" lastFinishedPulling="2025-11-24 19:56:53.939487024 +0000 UTC m=+2552.461993291" observedRunningTime="2025-11-24 19:56:54.49811464 +0000 UTC m=+2553.020620897" watchObservedRunningTime="2025-11-24 19:56:54.504368417 +0000 UTC m=+2553.026874684" Nov 24 19:56:59 crc kubenswrapper[5035]: I1124 19:56:59.523159 5035 generic.go:334] "Generic (PLEG): container finished" podID="6ef15736-c842-4a20-b05f-9105eef653a7" containerID="315d04c48f9fdbbea5c8dd5a0108decabe94e027cb666cb2ff91de3bbfbd18b0" exitCode=0 Nov 24 19:56:59 crc kubenswrapper[5035]: I1124 19:56:59.523276 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" event={"ID":"6ef15736-c842-4a20-b05f-9105eef653a7","Type":"ContainerDied","Data":"315d04c48f9fdbbea5c8dd5a0108decabe94e027cb666cb2ff91de3bbfbd18b0"} Nov 24 19:57:00 crc kubenswrapper[5035]: I1124 19:57:00.928534 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.028010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfmgn\" (UniqueName: \"kubernetes.io/projected/6ef15736-c842-4a20-b05f-9105eef653a7-kube-api-access-xfmgn\") pod \"6ef15736-c842-4a20-b05f-9105eef653a7\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.028348 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ceph\") pod \"6ef15736-c842-4a20-b05f-9105eef653a7\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.028476 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ssh-key\") pod \"6ef15736-c842-4a20-b05f-9105eef653a7\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.028621 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-inventory\") pod \"6ef15736-c842-4a20-b05f-9105eef653a7\" (UID: \"6ef15736-c842-4a20-b05f-9105eef653a7\") " Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.035421 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ef15736-c842-4a20-b05f-9105eef653a7-kube-api-access-xfmgn" (OuterVolumeSpecName: "kube-api-access-xfmgn") pod "6ef15736-c842-4a20-b05f-9105eef653a7" (UID: "6ef15736-c842-4a20-b05f-9105eef653a7"). InnerVolumeSpecName "kube-api-access-xfmgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.035741 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ceph" (OuterVolumeSpecName: "ceph") pod "6ef15736-c842-4a20-b05f-9105eef653a7" (UID: "6ef15736-c842-4a20-b05f-9105eef653a7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.066462 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6ef15736-c842-4a20-b05f-9105eef653a7" (UID: "6ef15736-c842-4a20-b05f-9105eef653a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.076444 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-inventory" (OuterVolumeSpecName: "inventory") pod "6ef15736-c842-4a20-b05f-9105eef653a7" (UID: "6ef15736-c842-4a20-b05f-9105eef653a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.130802 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfmgn\" (UniqueName: \"kubernetes.io/projected/6ef15736-c842-4a20-b05f-9105eef653a7-kube-api-access-xfmgn\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.130843 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.130852 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.130861 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef15736-c842-4a20-b05f-9105eef653a7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.542816 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" event={"ID":"6ef15736-c842-4a20-b05f-9105eef653a7","Type":"ContainerDied","Data":"29db042fe359fbfcdeed0fea14a02cb45548a71691b6533d8113ff67c8011e87"} Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.542873 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29db042fe359fbfcdeed0fea14a02cb45548a71691b6533d8113ff67c8011e87" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.542837 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.631762 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn"] Nov 24 19:57:01 crc kubenswrapper[5035]: E1124 19:57:01.632174 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef15736-c842-4a20-b05f-9105eef653a7" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.632195 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef15736-c842-4a20-b05f-9105eef653a7" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.632454 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef15736-c842-4a20-b05f-9105eef653a7" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.633066 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.635244 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.635270 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.635339 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.635930 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.636995 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.655433 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn"] Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.740180 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.740228 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmpmx\" (UniqueName: \"kubernetes.io/projected/d0ebd40d-a3e7-4358-93be-983c52b4f89d-kube-api-access-gmpmx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.740287 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.740358 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.841753 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.841814 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmpmx\" (UniqueName: \"kubernetes.io/projected/d0ebd40d-a3e7-4358-93be-983c52b4f89d-kube-api-access-gmpmx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.841869 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.841908 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.847880 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.848029 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.849235 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.864717 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmpmx\" (UniqueName: \"kubernetes.io/projected/d0ebd40d-a3e7-4358-93be-983c52b4f89d-kube-api-access-gmpmx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:01 crc kubenswrapper[5035]: I1124 19:57:01.953861 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:02 crc kubenswrapper[5035]: I1124 19:57:02.506083 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn"] Nov 24 19:57:02 crc kubenswrapper[5035]: W1124 19:57:02.515003 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0ebd40d_a3e7_4358_93be_983c52b4f89d.slice/crio-ce3f79a90483828524f9229c9a8b3b28a7b7e252242465d021ac92ee172ec7c9 WatchSource:0}: Error finding container ce3f79a90483828524f9229c9a8b3b28a7b7e252242465d021ac92ee172ec7c9: Status 404 returned error can't find the container with id ce3f79a90483828524f9229c9a8b3b28a7b7e252242465d021ac92ee172ec7c9 Nov 24 19:57:02 crc kubenswrapper[5035]: I1124 19:57:02.565441 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" event={"ID":"d0ebd40d-a3e7-4358-93be-983c52b4f89d","Type":"ContainerStarted","Data":"ce3f79a90483828524f9229c9a8b3b28a7b7e252242465d021ac92ee172ec7c9"} Nov 24 19:57:03 crc kubenswrapper[5035]: I1124 19:57:03.574710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" event={"ID":"d0ebd40d-a3e7-4358-93be-983c52b4f89d","Type":"ContainerStarted","Data":"0829b7475152310260065825a8cf501de2e372621d1c63ed0831f7a0d7ee154c"} Nov 24 19:57:03 crc kubenswrapper[5035]: I1124 19:57:03.594580 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" podStartSLOduration=2.182227051 podStartE2EDuration="2.594564381s" podCreationTimestamp="2025-11-24 19:57:01 +0000 UTC" firstStartedPulling="2025-11-24 19:57:02.517258793 +0000 UTC m=+2561.039765070" lastFinishedPulling="2025-11-24 19:57:02.929596143 +0000 UTC m=+2561.452102400" observedRunningTime="2025-11-24 19:57:03.591834638 +0000 UTC m=+2562.114340895" watchObservedRunningTime="2025-11-24 19:57:03.594564381 +0000 UTC m=+2562.117070628" Nov 24 19:57:55 crc kubenswrapper[5035]: I1124 19:57:55.065537 5035 generic.go:334] "Generic (PLEG): container finished" podID="d0ebd40d-a3e7-4358-93be-983c52b4f89d" containerID="0829b7475152310260065825a8cf501de2e372621d1c63ed0831f7a0d7ee154c" exitCode=0 Nov 24 19:57:55 crc kubenswrapper[5035]: I1124 19:57:55.065652 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" event={"ID":"d0ebd40d-a3e7-4358-93be-983c52b4f89d","Type":"ContainerDied","Data":"0829b7475152310260065825a8cf501de2e372621d1c63ed0831f7a0d7ee154c"} Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.482981 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.562614 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-inventory\") pod \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.562756 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ceph\") pod \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.562908 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ssh-key\") pod \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.563035 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmpmx\" (UniqueName: \"kubernetes.io/projected/d0ebd40d-a3e7-4358-93be-983c52b4f89d-kube-api-access-gmpmx\") pod \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\" (UID: \"d0ebd40d-a3e7-4358-93be-983c52b4f89d\") " Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.570306 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ebd40d-a3e7-4358-93be-983c52b4f89d-kube-api-access-gmpmx" (OuterVolumeSpecName: "kube-api-access-gmpmx") pod "d0ebd40d-a3e7-4358-93be-983c52b4f89d" (UID: "d0ebd40d-a3e7-4358-93be-983c52b4f89d"). InnerVolumeSpecName "kube-api-access-gmpmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.572327 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ceph" (OuterVolumeSpecName: "ceph") pod "d0ebd40d-a3e7-4358-93be-983c52b4f89d" (UID: "d0ebd40d-a3e7-4358-93be-983c52b4f89d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.595194 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d0ebd40d-a3e7-4358-93be-983c52b4f89d" (UID: "d0ebd40d-a3e7-4358-93be-983c52b4f89d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.597949 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-inventory" (OuterVolumeSpecName: "inventory") pod "d0ebd40d-a3e7-4358-93be-983c52b4f89d" (UID: "d0ebd40d-a3e7-4358-93be-983c52b4f89d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.664597 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.664642 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmpmx\" (UniqueName: \"kubernetes.io/projected/d0ebd40d-a3e7-4358-93be-983c52b4f89d-kube-api-access-gmpmx\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.664653 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:56 crc kubenswrapper[5035]: I1124 19:57:56.664663 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0ebd40d-a3e7-4358-93be-983c52b4f89d-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.086533 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" event={"ID":"d0ebd40d-a3e7-4358-93be-983c52b4f89d","Type":"ContainerDied","Data":"ce3f79a90483828524f9229c9a8b3b28a7b7e252242465d021ac92ee172ec7c9"} Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.086598 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce3f79a90483828524f9229c9a8b3b28a7b7e252242465d021ac92ee172ec7c9" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.087039 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.180497 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cnpqb"] Nov 24 19:57:57 crc kubenswrapper[5035]: E1124 19:57:57.181111 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ebd40d-a3e7-4358-93be-983c52b4f89d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.181132 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ebd40d-a3e7-4358-93be-983c52b4f89d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.181393 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ebd40d-a3e7-4358-93be-983c52b4f89d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.182011 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.184085 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.184562 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.184595 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.184617 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.184563 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.197705 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cnpqb"] Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.278523 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.278870 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ceph\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.279017 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4jmr\" (UniqueName: \"kubernetes.io/projected/f98da165-e7ef-4475-84db-01079d69bb82-kube-api-access-x4jmr\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.279091 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.381284 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.381502 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ceph\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.381615 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4jmr\" (UniqueName: \"kubernetes.io/projected/f98da165-e7ef-4475-84db-01079d69bb82-kube-api-access-x4jmr\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.381759 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.389254 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ceph\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.389427 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.392972 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.406899 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4jmr\" (UniqueName: \"kubernetes.io/projected/f98da165-e7ef-4475-84db-01079d69bb82-kube-api-access-x4jmr\") pod \"ssh-known-hosts-edpm-deployment-cnpqb\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:57 crc kubenswrapper[5035]: I1124 19:57:57.498015 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:57:58 crc kubenswrapper[5035]: I1124 19:57:58.024873 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cnpqb"] Nov 24 19:57:58 crc kubenswrapper[5035]: I1124 19:57:58.094553 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" event={"ID":"f98da165-e7ef-4475-84db-01079d69bb82","Type":"ContainerStarted","Data":"78f29abf2c43a625f12956e8c7abcb1333f66d4d5e156892700bc2224362c008"} Nov 24 19:57:59 crc kubenswrapper[5035]: I1124 19:57:59.104901 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" event={"ID":"f98da165-e7ef-4475-84db-01079d69bb82","Type":"ContainerStarted","Data":"b168fef5b2179fd254cd1edebf59a7693896fc8ce6eb3e269df70e5074edf098"} Nov 24 19:57:59 crc kubenswrapper[5035]: I1124 19:57:59.129251 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" podStartSLOduration=1.601565728 podStartE2EDuration="2.129228869s" podCreationTimestamp="2025-11-24 19:57:57 +0000 UTC" firstStartedPulling="2025-11-24 19:57:58.031481645 +0000 UTC m=+2616.553987902" lastFinishedPulling="2025-11-24 19:57:58.559144796 +0000 UTC m=+2617.081651043" observedRunningTime="2025-11-24 19:57:59.119259242 +0000 UTC m=+2617.641765509" watchObservedRunningTime="2025-11-24 19:57:59.129228869 +0000 UTC m=+2617.651735146" Nov 24 19:58:10 crc kubenswrapper[5035]: I1124 19:58:10.196638 5035 generic.go:334] "Generic (PLEG): container finished" podID="f98da165-e7ef-4475-84db-01079d69bb82" containerID="b168fef5b2179fd254cd1edebf59a7693896fc8ce6eb3e269df70e5074edf098" exitCode=0 Nov 24 19:58:10 crc kubenswrapper[5035]: I1124 19:58:10.196736 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" event={"ID":"f98da165-e7ef-4475-84db-01079d69bb82","Type":"ContainerDied","Data":"b168fef5b2179fd254cd1edebf59a7693896fc8ce6eb3e269df70e5074edf098"} Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.690200 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.883841 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-inventory-0\") pod \"f98da165-e7ef-4475-84db-01079d69bb82\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.884006 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ssh-key-openstack-edpm-ipam\") pod \"f98da165-e7ef-4475-84db-01079d69bb82\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.884156 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4jmr\" (UniqueName: \"kubernetes.io/projected/f98da165-e7ef-4475-84db-01079d69bb82-kube-api-access-x4jmr\") pod \"f98da165-e7ef-4475-84db-01079d69bb82\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.884251 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ceph\") pod \"f98da165-e7ef-4475-84db-01079d69bb82\" (UID: \"f98da165-e7ef-4475-84db-01079d69bb82\") " Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.889615 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ceph" (OuterVolumeSpecName: "ceph") pod "f98da165-e7ef-4475-84db-01079d69bb82" (UID: "f98da165-e7ef-4475-84db-01079d69bb82"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.890882 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f98da165-e7ef-4475-84db-01079d69bb82-kube-api-access-x4jmr" (OuterVolumeSpecName: "kube-api-access-x4jmr") pod "f98da165-e7ef-4475-84db-01079d69bb82" (UID: "f98da165-e7ef-4475-84db-01079d69bb82"). InnerVolumeSpecName "kube-api-access-x4jmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.915537 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f98da165-e7ef-4475-84db-01079d69bb82" (UID: "f98da165-e7ef-4475-84db-01079d69bb82"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.930671 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "f98da165-e7ef-4475-84db-01079d69bb82" (UID: "f98da165-e7ef-4475-84db-01079d69bb82"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.986007 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4jmr\" (UniqueName: \"kubernetes.io/projected/f98da165-e7ef-4475-84db-01079d69bb82-kube-api-access-x4jmr\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.986043 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.986057 5035 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:11 crc kubenswrapper[5035]: I1124 19:58:11.986070 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f98da165-e7ef-4475-84db-01079d69bb82-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.222089 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.225533 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cnpqb" event={"ID":"f98da165-e7ef-4475-84db-01079d69bb82","Type":"ContainerDied","Data":"78f29abf2c43a625f12956e8c7abcb1333f66d4d5e156892700bc2224362c008"} Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.225589 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f29abf2c43a625f12956e8c7abcb1333f66d4d5e156892700bc2224362c008" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.342644 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl"] Nov 24 19:58:12 crc kubenswrapper[5035]: E1124 19:58:12.343262 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98da165-e7ef-4475-84db-01079d69bb82" containerName="ssh-known-hosts-edpm-deployment" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.343394 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98da165-e7ef-4475-84db-01079d69bb82" containerName="ssh-known-hosts-edpm-deployment" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.343696 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f98da165-e7ef-4475-84db-01079d69bb82" containerName="ssh-known-hosts-edpm-deployment" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.344459 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.350937 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl"] Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.354347 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.354615 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.355078 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.355250 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.355437 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.392882 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.393182 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj8mj\" (UniqueName: \"kubernetes.io/projected/234d0af6-f029-4b29-8003-b954e9e4e07a-kube-api-access-hj8mj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.393418 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.393591 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.495066 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.495394 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj8mj\" (UniqueName: \"kubernetes.io/projected/234d0af6-f029-4b29-8003-b954e9e4e07a-kube-api-access-hj8mj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.495508 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.495612 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.499313 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.499357 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.500784 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.515638 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj8mj\" (UniqueName: \"kubernetes.io/projected/234d0af6-f029-4b29-8003-b954e9e4e07a-kube-api-access-hj8mj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kshfl\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:12 crc kubenswrapper[5035]: I1124 19:58:12.662999 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:13 crc kubenswrapper[5035]: I1124 19:58:13.258164 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl"] Nov 24 19:58:14 crc kubenswrapper[5035]: I1124 19:58:14.243928 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" event={"ID":"234d0af6-f029-4b29-8003-b954e9e4e07a","Type":"ContainerStarted","Data":"eff22613309fc1ce21fc57a8bd112a07268fb2ed56016d0ed156ad567deac6a7"} Nov 24 19:58:14 crc kubenswrapper[5035]: I1124 19:58:14.244475 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" event={"ID":"234d0af6-f029-4b29-8003-b954e9e4e07a","Type":"ContainerStarted","Data":"41433d4e3bbf023b5ee4f21e3ed75ef40d0cda74ed3696faa95600063d98a07b"} Nov 24 19:58:14 crc kubenswrapper[5035]: I1124 19:58:14.270122 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" podStartSLOduration=1.860522046 podStartE2EDuration="2.270092072s" podCreationTimestamp="2025-11-24 19:58:12 +0000 UTC" firstStartedPulling="2025-11-24 19:58:13.269581389 +0000 UTC m=+2631.792087656" lastFinishedPulling="2025-11-24 19:58:13.679151415 +0000 UTC m=+2632.201657682" observedRunningTime="2025-11-24 19:58:14.259245524 +0000 UTC m=+2632.781751811" watchObservedRunningTime="2025-11-24 19:58:14.270092072 +0000 UTC m=+2632.792598379" Nov 24 19:58:23 crc kubenswrapper[5035]: I1124 19:58:23.352872 5035 generic.go:334] "Generic (PLEG): container finished" podID="234d0af6-f029-4b29-8003-b954e9e4e07a" containerID="eff22613309fc1ce21fc57a8bd112a07268fb2ed56016d0ed156ad567deac6a7" exitCode=0 Nov 24 19:58:23 crc kubenswrapper[5035]: I1124 19:58:23.353770 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" event={"ID":"234d0af6-f029-4b29-8003-b954e9e4e07a","Type":"ContainerDied","Data":"eff22613309fc1ce21fc57a8bd112a07268fb2ed56016d0ed156ad567deac6a7"} Nov 24 19:58:24 crc kubenswrapper[5035]: I1124 19:58:24.781146 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:24 crc kubenswrapper[5035]: I1124 19:58:24.948239 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ceph\") pod \"234d0af6-f029-4b29-8003-b954e9e4e07a\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " Nov 24 19:58:24 crc kubenswrapper[5035]: I1124 19:58:24.948361 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj8mj\" (UniqueName: \"kubernetes.io/projected/234d0af6-f029-4b29-8003-b954e9e4e07a-kube-api-access-hj8mj\") pod \"234d0af6-f029-4b29-8003-b954e9e4e07a\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " Nov 24 19:58:24 crc kubenswrapper[5035]: I1124 19:58:24.948444 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ssh-key\") pod \"234d0af6-f029-4b29-8003-b954e9e4e07a\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " Nov 24 19:58:24 crc kubenswrapper[5035]: I1124 19:58:24.948608 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-inventory\") pod \"234d0af6-f029-4b29-8003-b954e9e4e07a\" (UID: \"234d0af6-f029-4b29-8003-b954e9e4e07a\") " Nov 24 19:58:24 crc kubenswrapper[5035]: I1124 19:58:24.972018 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ceph" (OuterVolumeSpecName: "ceph") pod "234d0af6-f029-4b29-8003-b954e9e4e07a" (UID: "234d0af6-f029-4b29-8003-b954e9e4e07a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.001525 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234d0af6-f029-4b29-8003-b954e9e4e07a-kube-api-access-hj8mj" (OuterVolumeSpecName: "kube-api-access-hj8mj") pod "234d0af6-f029-4b29-8003-b954e9e4e07a" (UID: "234d0af6-f029-4b29-8003-b954e9e4e07a"). InnerVolumeSpecName "kube-api-access-hj8mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.021551 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-inventory" (OuterVolumeSpecName: "inventory") pod "234d0af6-f029-4b29-8003-b954e9e4e07a" (UID: "234d0af6-f029-4b29-8003-b954e9e4e07a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.050646 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.050688 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj8mj\" (UniqueName: \"kubernetes.io/projected/234d0af6-f029-4b29-8003-b954e9e4e07a-kube-api-access-hj8mj\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.050753 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.074465 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "234d0af6-f029-4b29-8003-b954e9e4e07a" (UID: "234d0af6-f029-4b29-8003-b954e9e4e07a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.152782 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/234d0af6-f029-4b29-8003-b954e9e4e07a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.373651 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" event={"ID":"234d0af6-f029-4b29-8003-b954e9e4e07a","Type":"ContainerDied","Data":"41433d4e3bbf023b5ee4f21e3ed75ef40d0cda74ed3696faa95600063d98a07b"} Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.373683 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41433d4e3bbf023b5ee4f21e3ed75ef40d0cda74ed3696faa95600063d98a07b" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.373742 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kshfl" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.461793 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd"] Nov 24 19:58:25 crc kubenswrapper[5035]: E1124 19:58:25.462248 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234d0af6-f029-4b29-8003-b954e9e4e07a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.462270 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="234d0af6-f029-4b29-8003-b954e9e4e07a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.462477 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="234d0af6-f029-4b29-8003-b954e9e4e07a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.463126 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.468218 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.468225 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.468458 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.468766 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.468826 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.470784 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd"] Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.559192 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.559251 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q8zp\" (UniqueName: \"kubernetes.io/projected/31438de3-4165-4a60-8a13-bb24b1743cd6-kube-api-access-7q8zp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.559402 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.559432 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.661119 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.661233 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.661276 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q8zp\" (UniqueName: \"kubernetes.io/projected/31438de3-4165-4a60-8a13-bb24b1743cd6-kube-api-access-7q8zp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.661401 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.667449 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.667508 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.668747 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.680365 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q8zp\" (UniqueName: \"kubernetes.io/projected/31438de3-4165-4a60-8a13-bb24b1743cd6-kube-api-access-7q8zp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:25 crc kubenswrapper[5035]: I1124 19:58:25.790112 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:26 crc kubenswrapper[5035]: I1124 19:58:26.400257 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" event={"ID":"31438de3-4165-4a60-8a13-bb24b1743cd6","Type":"ContainerStarted","Data":"a82a65e378744db0d9303a9d897c5c2d24b9e0d0384a24374e2c9b89a1cbe2c1"} Nov 24 19:58:26 crc kubenswrapper[5035]: I1124 19:58:26.401536 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd"] Nov 24 19:58:27 crc kubenswrapper[5035]: I1124 19:58:27.410772 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" event={"ID":"31438de3-4165-4a60-8a13-bb24b1743cd6","Type":"ContainerStarted","Data":"8d215f1f77aa6ccbbd90aa5737485756e69ad79b7ef1c4843b2d776d49d38a0f"} Nov 24 19:58:27 crc kubenswrapper[5035]: I1124 19:58:27.431171 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" podStartSLOduration=1.998580024 podStartE2EDuration="2.431149244s" podCreationTimestamp="2025-11-24 19:58:25 +0000 UTC" firstStartedPulling="2025-11-24 19:58:26.380946207 +0000 UTC m=+2644.903452464" lastFinishedPulling="2025-11-24 19:58:26.813515417 +0000 UTC m=+2645.336021684" observedRunningTime="2025-11-24 19:58:27.427704083 +0000 UTC m=+2645.950210330" watchObservedRunningTime="2025-11-24 19:58:27.431149244 +0000 UTC m=+2645.953655501" Nov 24 19:58:38 crc kubenswrapper[5035]: I1124 19:58:38.509817 5035 generic.go:334] "Generic (PLEG): container finished" podID="31438de3-4165-4a60-8a13-bb24b1743cd6" containerID="8d215f1f77aa6ccbbd90aa5737485756e69ad79b7ef1c4843b2d776d49d38a0f" exitCode=0 Nov 24 19:58:38 crc kubenswrapper[5035]: I1124 19:58:38.509939 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" event={"ID":"31438de3-4165-4a60-8a13-bb24b1743cd6","Type":"ContainerDied","Data":"8d215f1f77aa6ccbbd90aa5737485756e69ad79b7ef1c4843b2d776d49d38a0f"} Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.385125 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v8zql"] Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.388406 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.405134 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v8zql"] Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.442702 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-utilities\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.442858 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-catalog-content\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.443035 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82mj4\" (UniqueName: \"kubernetes.io/projected/10ff1a01-fa3e-4fb8-87fb-0934df48619d-kube-api-access-82mj4\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.545353 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-utilities\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.545450 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-catalog-content\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.545575 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82mj4\" (UniqueName: \"kubernetes.io/projected/10ff1a01-fa3e-4fb8-87fb-0934df48619d-kube-api-access-82mj4\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.545990 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-utilities\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.546133 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-catalog-content\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.568073 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82mj4\" (UniqueName: \"kubernetes.io/projected/10ff1a01-fa3e-4fb8-87fb-0934df48619d-kube-api-access-82mj4\") pod \"redhat-operators-v8zql\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:39 crc kubenswrapper[5035]: I1124 19:58:39.711849 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.026385 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.156028 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ceph\") pod \"31438de3-4165-4a60-8a13-bb24b1743cd6\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.156283 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ssh-key\") pod \"31438de3-4165-4a60-8a13-bb24b1743cd6\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.156404 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-inventory\") pod \"31438de3-4165-4a60-8a13-bb24b1743cd6\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.156503 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q8zp\" (UniqueName: \"kubernetes.io/projected/31438de3-4165-4a60-8a13-bb24b1743cd6-kube-api-access-7q8zp\") pod \"31438de3-4165-4a60-8a13-bb24b1743cd6\" (UID: \"31438de3-4165-4a60-8a13-bb24b1743cd6\") " Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.165850 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ceph" (OuterVolumeSpecName: "ceph") pod "31438de3-4165-4a60-8a13-bb24b1743cd6" (UID: "31438de3-4165-4a60-8a13-bb24b1743cd6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.169571 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31438de3-4165-4a60-8a13-bb24b1743cd6-kube-api-access-7q8zp" (OuterVolumeSpecName: "kube-api-access-7q8zp") pod "31438de3-4165-4a60-8a13-bb24b1743cd6" (UID: "31438de3-4165-4a60-8a13-bb24b1743cd6"). InnerVolumeSpecName "kube-api-access-7q8zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.190645 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-inventory" (OuterVolumeSpecName: "inventory") pod "31438de3-4165-4a60-8a13-bb24b1743cd6" (UID: "31438de3-4165-4a60-8a13-bb24b1743cd6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.191326 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31438de3-4165-4a60-8a13-bb24b1743cd6" (UID: "31438de3-4165-4a60-8a13-bb24b1743cd6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.250274 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v8zql"] Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.258524 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q8zp\" (UniqueName: \"kubernetes.io/projected/31438de3-4165-4a60-8a13-bb24b1743cd6-kube-api-access-7q8zp\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.258556 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.258566 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.258575 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31438de3-4165-4a60-8a13-bb24b1743cd6-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.530937 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" event={"ID":"31438de3-4165-4a60-8a13-bb24b1743cd6","Type":"ContainerDied","Data":"a82a65e378744db0d9303a9d897c5c2d24b9e0d0384a24374e2c9b89a1cbe2c1"} Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.531173 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a82a65e378744db0d9303a9d897c5c2d24b9e0d0384a24374e2c9b89a1cbe2c1" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.530963 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.533007 5035 generic.go:334] "Generic (PLEG): container finished" podID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerID="7c0ac2bbd94c8912cd1825f53bda03c2cda489ad629d6751f65f0ba95eb733e6" exitCode=0 Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.533051 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8zql" event={"ID":"10ff1a01-fa3e-4fb8-87fb-0934df48619d","Type":"ContainerDied","Data":"7c0ac2bbd94c8912cd1825f53bda03c2cda489ad629d6751f65f0ba95eb733e6"} Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.533077 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8zql" event={"ID":"10ff1a01-fa3e-4fb8-87fb-0934df48619d","Type":"ContainerStarted","Data":"a5701a71f35a0604c876c27471d23765039d90a025eeadf6785cf6b271e323c5"} Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.630187 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4"] Nov 24 19:58:40 crc kubenswrapper[5035]: E1124 19:58:40.630946 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31438de3-4165-4a60-8a13-bb24b1743cd6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.630964 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="31438de3-4165-4a60-8a13-bb24b1743cd6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.631209 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="31438de3-4165-4a60-8a13-bb24b1743cd6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.632047 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.635240 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.635440 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.635619 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.635739 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.638860 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.639055 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.639444 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.639746 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.648645 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4"] Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775268 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775332 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775355 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775391 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775415 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775433 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775456 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9295c\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-kube-api-access-9295c\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775479 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775524 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775656 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775719 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775853 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.775984 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877377 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877400 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877417 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877439 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877468 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877487 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877504 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9295c\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-kube-api-access-9295c\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877523 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877567 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877597 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877617 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.877640 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.883780 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.884254 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.884568 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.884735 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.884928 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.885238 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.885942 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.887318 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.887875 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.888624 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.890490 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.891133 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.898777 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9295c\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-kube-api-access-9295c\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:40 crc kubenswrapper[5035]: I1124 19:58:40.977928 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:58:41 crc kubenswrapper[5035]: I1124 19:58:41.533592 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4"] Nov 24 19:58:41 crc kubenswrapper[5035]: W1124 19:58:41.543106 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbae54c1b_690d_4e84_8c7e_6f4fdab2ae26.slice/crio-f952654cd2c99c4cae99a7049cce3816ccefdd792020c22da66761497f106014 WatchSource:0}: Error finding container f952654cd2c99c4cae99a7049cce3816ccefdd792020c22da66761497f106014: Status 404 returned error can't find the container with id f952654cd2c99c4cae99a7049cce3816ccefdd792020c22da66761497f106014 Nov 24 19:58:42 crc kubenswrapper[5035]: I1124 19:58:42.556563 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" event={"ID":"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26","Type":"ContainerStarted","Data":"51c6a6cbe70ff1058d47d099cb7860d78263e85375502a72e2b24c1f524c6960"} Nov 24 19:58:42 crc kubenswrapper[5035]: I1124 19:58:42.556784 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" event={"ID":"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26","Type":"ContainerStarted","Data":"f952654cd2c99c4cae99a7049cce3816ccefdd792020c22da66761497f106014"} Nov 24 19:58:42 crc kubenswrapper[5035]: I1124 19:58:42.559488 5035 generic.go:334] "Generic (PLEG): container finished" podID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerID="82354ba8a3949cfc0b8e8ea8f7d39a1628d4a5219a30059a382e1e1d98cb1237" exitCode=0 Nov 24 19:58:42 crc kubenswrapper[5035]: I1124 19:58:42.559552 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8zql" event={"ID":"10ff1a01-fa3e-4fb8-87fb-0934df48619d","Type":"ContainerDied","Data":"82354ba8a3949cfc0b8e8ea8f7d39a1628d4a5219a30059a382e1e1d98cb1237"} Nov 24 19:58:42 crc kubenswrapper[5035]: I1124 19:58:42.586116 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" podStartSLOduration=2.079634258 podStartE2EDuration="2.586095926s" podCreationTimestamp="2025-11-24 19:58:40 +0000 UTC" firstStartedPulling="2025-11-24 19:58:41.547890108 +0000 UTC m=+2660.070396355" lastFinishedPulling="2025-11-24 19:58:42.054351766 +0000 UTC m=+2660.576858023" observedRunningTime="2025-11-24 19:58:42.573345777 +0000 UTC m=+2661.095852034" watchObservedRunningTime="2025-11-24 19:58:42.586095926 +0000 UTC m=+2661.108602183" Nov 24 19:58:43 crc kubenswrapper[5035]: I1124 19:58:43.568234 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8zql" event={"ID":"10ff1a01-fa3e-4fb8-87fb-0934df48619d","Type":"ContainerStarted","Data":"b7826b097ffcced648b1b11d8b3320b09b472a03e1fa711ab3f387e38e0d25a7"} Nov 24 19:58:43 crc kubenswrapper[5035]: I1124 19:58:43.589490 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v8zql" podStartSLOduration=2.184835881 podStartE2EDuration="4.589471927s" podCreationTimestamp="2025-11-24 19:58:39 +0000 UTC" firstStartedPulling="2025-11-24 19:58:40.53480496 +0000 UTC m=+2659.057311207" lastFinishedPulling="2025-11-24 19:58:42.939440996 +0000 UTC m=+2661.461947253" observedRunningTime="2025-11-24 19:58:43.586050865 +0000 UTC m=+2662.108557122" watchObservedRunningTime="2025-11-24 19:58:43.589471927 +0000 UTC m=+2662.111978184" Nov 24 19:58:49 crc kubenswrapper[5035]: I1124 19:58:49.713221 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:49 crc kubenswrapper[5035]: I1124 19:58:49.713923 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:49 crc kubenswrapper[5035]: I1124 19:58:49.804449 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:50 crc kubenswrapper[5035]: I1124 19:58:50.680469 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:51 crc kubenswrapper[5035]: I1124 19:58:51.047964 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v8zql"] Nov 24 19:58:52 crc kubenswrapper[5035]: I1124 19:58:52.640365 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v8zql" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="registry-server" containerID="cri-o://b7826b097ffcced648b1b11d8b3320b09b472a03e1fa711ab3f387e38e0d25a7" gracePeriod=2 Nov 24 19:58:54 crc kubenswrapper[5035]: I1124 19:58:54.665584 5035 generic.go:334] "Generic (PLEG): container finished" podID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerID="b7826b097ffcced648b1b11d8b3320b09b472a03e1fa711ab3f387e38e0d25a7" exitCode=0 Nov 24 19:58:54 crc kubenswrapper[5035]: I1124 19:58:54.665672 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8zql" event={"ID":"10ff1a01-fa3e-4fb8-87fb-0934df48619d","Type":"ContainerDied","Data":"b7826b097ffcced648b1b11d8b3320b09b472a03e1fa711ab3f387e38e0d25a7"} Nov 24 19:58:54 crc kubenswrapper[5035]: I1124 19:58:54.887146 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.088773 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-utilities\") pod \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.088942 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-catalog-content\") pod \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.089049 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82mj4\" (UniqueName: \"kubernetes.io/projected/10ff1a01-fa3e-4fb8-87fb-0934df48619d-kube-api-access-82mj4\") pod \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\" (UID: \"10ff1a01-fa3e-4fb8-87fb-0934df48619d\") " Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.090425 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-utilities" (OuterVolumeSpecName: "utilities") pod "10ff1a01-fa3e-4fb8-87fb-0934df48619d" (UID: "10ff1a01-fa3e-4fb8-87fb-0934df48619d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.095018 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10ff1a01-fa3e-4fb8-87fb-0934df48619d-kube-api-access-82mj4" (OuterVolumeSpecName: "kube-api-access-82mj4") pod "10ff1a01-fa3e-4fb8-87fb-0934df48619d" (UID: "10ff1a01-fa3e-4fb8-87fb-0934df48619d"). InnerVolumeSpecName "kube-api-access-82mj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.191720 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.192483 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82mj4\" (UniqueName: \"kubernetes.io/projected/10ff1a01-fa3e-4fb8-87fb-0934df48619d-kube-api-access-82mj4\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.229246 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10ff1a01-fa3e-4fb8-87fb-0934df48619d" (UID: "10ff1a01-fa3e-4fb8-87fb-0934df48619d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.294577 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ff1a01-fa3e-4fb8-87fb-0934df48619d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.678585 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v8zql" event={"ID":"10ff1a01-fa3e-4fb8-87fb-0934df48619d","Type":"ContainerDied","Data":"a5701a71f35a0604c876c27471d23765039d90a025eeadf6785cf6b271e323c5"} Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.678639 5035 scope.go:117] "RemoveContainer" containerID="b7826b097ffcced648b1b11d8b3320b09b472a03e1fa711ab3f387e38e0d25a7" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.678655 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v8zql" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.718515 5035 scope.go:117] "RemoveContainer" containerID="82354ba8a3949cfc0b8e8ea8f7d39a1628d4a5219a30059a382e1e1d98cb1237" Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.729262 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v8zql"] Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.743071 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v8zql"] Nov 24 19:58:55 crc kubenswrapper[5035]: I1124 19:58:55.753872 5035 scope.go:117] "RemoveContainer" containerID="7c0ac2bbd94c8912cd1825f53bda03c2cda489ad629d6751f65f0ba95eb733e6" Nov 24 19:58:56 crc kubenswrapper[5035]: I1124 19:58:56.222074 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" path="/var/lib/kubelet/pods/10ff1a01-fa3e-4fb8-87fb-0934df48619d/volumes" Nov 24 19:59:15 crc kubenswrapper[5035]: I1124 19:59:15.233971 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:59:15 crc kubenswrapper[5035]: I1124 19:59:15.234737 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 19:59:20 crc kubenswrapper[5035]: I1124 19:59:20.933700 5035 generic.go:334] "Generic (PLEG): container finished" podID="bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" containerID="51c6a6cbe70ff1058d47d099cb7860d78263e85375502a72e2b24c1f524c6960" exitCode=0 Nov 24 19:59:20 crc kubenswrapper[5035]: I1124 19:59:20.933799 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" event={"ID":"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26","Type":"ContainerDied","Data":"51c6a6cbe70ff1058d47d099cb7860d78263e85375502a72e2b24c1f524c6960"} Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.418996 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.558825 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.558971 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-libvirt-combined-ca-bundle\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.559019 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-repo-setup-combined-ca-bundle\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.559116 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-nova-combined-ca-bundle\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560124 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-inventory\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560335 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9295c\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-kube-api-access-9295c\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560479 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-neutron-metadata-combined-ca-bundle\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560558 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-ovn-default-certs-0\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560633 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ssh-key\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560785 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560866 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-bootstrap-combined-ca-bundle\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.560919 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ovn-combined-ca-bundle\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.561045 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ceph\") pod \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\" (UID: \"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26\") " Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.564877 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.565337 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.565359 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.565442 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.565473 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.566831 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ceph" (OuterVolumeSpecName: "ceph") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.568479 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.568737 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-kube-api-access-9295c" (OuterVolumeSpecName: "kube-api-access-9295c") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "kube-api-access-9295c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.568754 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.571835 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.573523 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.587965 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.593635 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-inventory" (OuterVolumeSpecName: "inventory") pod "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" (UID: "bae54c1b-690d-4e84-8c7e-6f4fdab2ae26"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663837 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9295c\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-kube-api-access-9295c\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663871 5035 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663884 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663894 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663904 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663919 5035 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663930 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663940 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663950 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663961 5035 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663974 5035 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663986 5035 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.663998 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae54c1b-690d-4e84-8c7e-6f4fdab2ae26-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.955963 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" event={"ID":"bae54c1b-690d-4e84-8c7e-6f4fdab2ae26","Type":"ContainerDied","Data":"f952654cd2c99c4cae99a7049cce3816ccefdd792020c22da66761497f106014"} Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.956006 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f952654cd2c99c4cae99a7049cce3816ccefdd792020c22da66761497f106014" Nov 24 19:59:22 crc kubenswrapper[5035]: I1124 19:59:22.956088 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.055812 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb"] Nov 24 19:59:23 crc kubenswrapper[5035]: E1124 19:59:23.056188 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.056207 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 19:59:23 crc kubenswrapper[5035]: E1124 19:59:23.056223 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="extract-utilities" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.056229 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="extract-utilities" Nov 24 19:59:23 crc kubenswrapper[5035]: E1124 19:59:23.056240 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="extract-content" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.056245 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="extract-content" Nov 24 19:59:23 crc kubenswrapper[5035]: E1124 19:59:23.056260 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="registry-server" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.056265 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="registry-server" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.056468 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae54c1b-690d-4e84-8c7e-6f4fdab2ae26" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.056484 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ff1a01-fa3e-4fb8-87fb-0934df48619d" containerName="registry-server" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.057052 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.059559 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.059714 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.059736 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.059811 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.061819 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.077535 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb"] Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.172216 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j88c6\" (UniqueName: \"kubernetes.io/projected/749b308d-659f-4128-b53c-cefd8ab3e1d6-kube-api-access-j88c6\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.172269 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.172345 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.172371 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.274649 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j88c6\" (UniqueName: \"kubernetes.io/projected/749b308d-659f-4128-b53c-cefd8ab3e1d6-kube-api-access-j88c6\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.274773 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.274885 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.274938 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.280528 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.282600 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.282855 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.300695 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j88c6\" (UniqueName: \"kubernetes.io/projected/749b308d-659f-4128-b53c-cefd8ab3e1d6-kube-api-access-j88c6\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.378737 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.921363 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb"] Nov 24 19:59:23 crc kubenswrapper[5035]: W1124 19:59:23.927060 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod749b308d_659f_4128_b53c_cefd8ab3e1d6.slice/crio-25e3fb1049447e7957683693502e03bd3b54c6d1ce1563a63614f71677115c2d WatchSource:0}: Error finding container 25e3fb1049447e7957683693502e03bd3b54c6d1ce1563a63614f71677115c2d: Status 404 returned error can't find the container with id 25e3fb1049447e7957683693502e03bd3b54c6d1ce1563a63614f71677115c2d Nov 24 19:59:23 crc kubenswrapper[5035]: I1124 19:59:23.967557 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" event={"ID":"749b308d-659f-4128-b53c-cefd8ab3e1d6","Type":"ContainerStarted","Data":"25e3fb1049447e7957683693502e03bd3b54c6d1ce1563a63614f71677115c2d"} Nov 24 19:59:25 crc kubenswrapper[5035]: I1124 19:59:25.986836 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" event={"ID":"749b308d-659f-4128-b53c-cefd8ab3e1d6","Type":"ContainerStarted","Data":"d8683b02a39c3383788fac8e4317d069b7caba98c4d5eba0c7fb6e820bd5ca16"} Nov 24 19:59:26 crc kubenswrapper[5035]: I1124 19:59:26.006924 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" podStartSLOduration=2.193462726 podStartE2EDuration="3.006906237s" podCreationTimestamp="2025-11-24 19:59:23 +0000 UTC" firstStartedPulling="2025-11-24 19:59:23.931746885 +0000 UTC m=+2702.454253152" lastFinishedPulling="2025-11-24 19:59:24.745190406 +0000 UTC m=+2703.267696663" observedRunningTime="2025-11-24 19:59:26.005952341 +0000 UTC m=+2704.528458598" watchObservedRunningTime="2025-11-24 19:59:26.006906237 +0000 UTC m=+2704.529412494" Nov 24 19:59:32 crc kubenswrapper[5035]: I1124 19:59:32.045317 5035 generic.go:334] "Generic (PLEG): container finished" podID="749b308d-659f-4128-b53c-cefd8ab3e1d6" containerID="d8683b02a39c3383788fac8e4317d069b7caba98c4d5eba0c7fb6e820bd5ca16" exitCode=0 Nov 24 19:59:32 crc kubenswrapper[5035]: I1124 19:59:32.045395 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" event={"ID":"749b308d-659f-4128-b53c-cefd8ab3e1d6","Type":"ContainerDied","Data":"d8683b02a39c3383788fac8e4317d069b7caba98c4d5eba0c7fb6e820bd5ca16"} Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.480771 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.592853 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-inventory\") pod \"749b308d-659f-4128-b53c-cefd8ab3e1d6\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.592936 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ceph\") pod \"749b308d-659f-4128-b53c-cefd8ab3e1d6\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.593036 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j88c6\" (UniqueName: \"kubernetes.io/projected/749b308d-659f-4128-b53c-cefd8ab3e1d6-kube-api-access-j88c6\") pod \"749b308d-659f-4128-b53c-cefd8ab3e1d6\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.593061 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ssh-key\") pod \"749b308d-659f-4128-b53c-cefd8ab3e1d6\" (UID: \"749b308d-659f-4128-b53c-cefd8ab3e1d6\") " Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.599214 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ceph" (OuterVolumeSpecName: "ceph") pod "749b308d-659f-4128-b53c-cefd8ab3e1d6" (UID: "749b308d-659f-4128-b53c-cefd8ab3e1d6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.599255 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/749b308d-659f-4128-b53c-cefd8ab3e1d6-kube-api-access-j88c6" (OuterVolumeSpecName: "kube-api-access-j88c6") pod "749b308d-659f-4128-b53c-cefd8ab3e1d6" (UID: "749b308d-659f-4128-b53c-cefd8ab3e1d6"). InnerVolumeSpecName "kube-api-access-j88c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.630620 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "749b308d-659f-4128-b53c-cefd8ab3e1d6" (UID: "749b308d-659f-4128-b53c-cefd8ab3e1d6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.630991 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-inventory" (OuterVolumeSpecName: "inventory") pod "749b308d-659f-4128-b53c-cefd8ab3e1d6" (UID: "749b308d-659f-4128-b53c-cefd8ab3e1d6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.694633 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.694675 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j88c6\" (UniqueName: \"kubernetes.io/projected/749b308d-659f-4128-b53c-cefd8ab3e1d6-kube-api-access-j88c6\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.694696 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:33 crc kubenswrapper[5035]: I1124 19:59:33.694710 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/749b308d-659f-4128-b53c-cefd8ab3e1d6-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.077918 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" event={"ID":"749b308d-659f-4128-b53c-cefd8ab3e1d6","Type":"ContainerDied","Data":"25e3fb1049447e7957683693502e03bd3b54c6d1ce1563a63614f71677115c2d"} Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.078420 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25e3fb1049447e7957683693502e03bd3b54c6d1ce1563a63614f71677115c2d" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.078001 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.157761 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz"] Nov 24 19:59:34 crc kubenswrapper[5035]: E1124 19:59:34.158204 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="749b308d-659f-4128-b53c-cefd8ab3e1d6" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.158225 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="749b308d-659f-4128-b53c-cefd8ab3e1d6" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.160672 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="749b308d-659f-4128-b53c-cefd8ab3e1d6" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.161442 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.165450 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.165628 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.165753 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.168120 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.168502 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.168613 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.175746 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz"] Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.306729 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.307052 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pktlh\" (UniqueName: \"kubernetes.io/projected/f2273d9c-4343-4bb5-9bca-0a87c601cf15-kube-api-access-pktlh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.307281 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.307812 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.307894 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.308020 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.410141 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.410262 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.410310 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.410354 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.410408 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.410445 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pktlh\" (UniqueName: \"kubernetes.io/projected/f2273d9c-4343-4bb5-9bca-0a87c601cf15-kube-api-access-pktlh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.411823 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.415598 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.416708 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.419868 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.421075 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.438029 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pktlh\" (UniqueName: \"kubernetes.io/projected/f2273d9c-4343-4bb5-9bca-0a87c601cf15-kube-api-access-pktlh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-b4rlz\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.487839 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 19:59:34 crc kubenswrapper[5035]: I1124 19:59:34.998369 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz"] Nov 24 19:59:35 crc kubenswrapper[5035]: I1124 19:59:35.088001 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" event={"ID":"f2273d9c-4343-4bb5-9bca-0a87c601cf15","Type":"ContainerStarted","Data":"ff3b88161d157c558fd88c4832632cbf9c68295e28eeac6b56f1c6f907e41558"} Nov 24 19:59:36 crc kubenswrapper[5035]: I1124 19:59:36.108532 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" event={"ID":"f2273d9c-4343-4bb5-9bca-0a87c601cf15","Type":"ContainerStarted","Data":"9da177310c63d9d083b4288395c6548859d31059e826eac6335e170c5d099930"} Nov 24 19:59:36 crc kubenswrapper[5035]: I1124 19:59:36.139773 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" podStartSLOduration=1.668357525 podStartE2EDuration="2.139745718s" podCreationTimestamp="2025-11-24 19:59:34 +0000 UTC" firstStartedPulling="2025-11-24 19:59:35.004618239 +0000 UTC m=+2713.527124496" lastFinishedPulling="2025-11-24 19:59:35.476006392 +0000 UTC m=+2713.998512689" observedRunningTime="2025-11-24 19:59:36.139276685 +0000 UTC m=+2714.661782972" watchObservedRunningTime="2025-11-24 19:59:36.139745718 +0000 UTC m=+2714.662251975" Nov 24 19:59:45 crc kubenswrapper[5035]: I1124 19:59:45.236684 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 19:59:45 crc kubenswrapper[5035]: I1124 19:59:45.237337 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.154851 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv"] Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.156533 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.158472 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.159413 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.168869 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv"] Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.260538 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72106bf6-5bff-4412-b58e-47c869df5338-config-volume\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.260740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wxr4\" (UniqueName: \"kubernetes.io/projected/72106bf6-5bff-4412-b58e-47c869df5338-kube-api-access-7wxr4\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.260766 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72106bf6-5bff-4412-b58e-47c869df5338-secret-volume\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.362158 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wxr4\" (UniqueName: \"kubernetes.io/projected/72106bf6-5bff-4412-b58e-47c869df5338-kube-api-access-7wxr4\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.362268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72106bf6-5bff-4412-b58e-47c869df5338-secret-volume\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.363246 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72106bf6-5bff-4412-b58e-47c869df5338-config-volume\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.363951 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72106bf6-5bff-4412-b58e-47c869df5338-config-volume\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.370689 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72106bf6-5bff-4412-b58e-47c869df5338-secret-volume\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.382233 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wxr4\" (UniqueName: \"kubernetes.io/projected/72106bf6-5bff-4412-b58e-47c869df5338-kube-api-access-7wxr4\") pod \"collect-profiles-29400240-tprvv\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.476147 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:00 crc kubenswrapper[5035]: I1124 20:00:00.914773 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv"] Nov 24 20:00:01 crc kubenswrapper[5035]: I1124 20:00:01.364655 5035 generic.go:334] "Generic (PLEG): container finished" podID="72106bf6-5bff-4412-b58e-47c869df5338" containerID="c4e87ae1e98c69667633be750652a0ce9517239f1e37682e94a4679d23ef1f40" exitCode=0 Nov 24 20:00:01 crc kubenswrapper[5035]: I1124 20:00:01.364817 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" event={"ID":"72106bf6-5bff-4412-b58e-47c869df5338","Type":"ContainerDied","Data":"c4e87ae1e98c69667633be750652a0ce9517239f1e37682e94a4679d23ef1f40"} Nov 24 20:00:01 crc kubenswrapper[5035]: I1124 20:00:01.364953 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" event={"ID":"72106bf6-5bff-4412-b58e-47c869df5338","Type":"ContainerStarted","Data":"064c814dd32d3587480c806f6afadd4ac588801331e7505aae902c4525a6f94b"} Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.695088 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.817815 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wxr4\" (UniqueName: \"kubernetes.io/projected/72106bf6-5bff-4412-b58e-47c869df5338-kube-api-access-7wxr4\") pod \"72106bf6-5bff-4412-b58e-47c869df5338\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.818199 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72106bf6-5bff-4412-b58e-47c869df5338-secret-volume\") pod \"72106bf6-5bff-4412-b58e-47c869df5338\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.818348 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72106bf6-5bff-4412-b58e-47c869df5338-config-volume\") pod \"72106bf6-5bff-4412-b58e-47c869df5338\" (UID: \"72106bf6-5bff-4412-b58e-47c869df5338\") " Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.818928 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72106bf6-5bff-4412-b58e-47c869df5338-config-volume" (OuterVolumeSpecName: "config-volume") pod "72106bf6-5bff-4412-b58e-47c869df5338" (UID: "72106bf6-5bff-4412-b58e-47c869df5338"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.824796 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72106bf6-5bff-4412-b58e-47c869df5338-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "72106bf6-5bff-4412-b58e-47c869df5338" (UID: "72106bf6-5bff-4412-b58e-47c869df5338"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.824923 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72106bf6-5bff-4412-b58e-47c869df5338-kube-api-access-7wxr4" (OuterVolumeSpecName: "kube-api-access-7wxr4") pod "72106bf6-5bff-4412-b58e-47c869df5338" (UID: "72106bf6-5bff-4412-b58e-47c869df5338"). InnerVolumeSpecName "kube-api-access-7wxr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.919462 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wxr4\" (UniqueName: \"kubernetes.io/projected/72106bf6-5bff-4412-b58e-47c869df5338-kube-api-access-7wxr4\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.919526 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/72106bf6-5bff-4412-b58e-47c869df5338-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:02 crc kubenswrapper[5035]: I1124 20:00:02.919543 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/72106bf6-5bff-4412-b58e-47c869df5338-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:03 crc kubenswrapper[5035]: I1124 20:00:03.388459 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" event={"ID":"72106bf6-5bff-4412-b58e-47c869df5338","Type":"ContainerDied","Data":"064c814dd32d3587480c806f6afadd4ac588801331e7505aae902c4525a6f94b"} Nov 24 20:00:03 crc kubenswrapper[5035]: I1124 20:00:03.388516 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="064c814dd32d3587480c806f6afadd4ac588801331e7505aae902c4525a6f94b" Nov 24 20:00:03 crc kubenswrapper[5035]: I1124 20:00:03.388565 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv" Nov 24 20:00:03 crc kubenswrapper[5035]: I1124 20:00:03.785825 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn"] Nov 24 20:00:03 crc kubenswrapper[5035]: I1124 20:00:03.792869 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400195-fd2dn"] Nov 24 20:00:04 crc kubenswrapper[5035]: I1124 20:00:04.216396 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d4ae5fb-edf2-4b2b-b462-60146b3e7a88" path="/var/lib/kubelet/pods/0d4ae5fb-edf2-4b2b-b462-60146b3e7a88/volumes" Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.234003 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.236941 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.237136 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.238573 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25c3576c8f6c0c703275007665c285ca572de78d4e00d849bb0e88dffe634030"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.238776 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://25c3576c8f6c0c703275007665c285ca572de78d4e00d849bb0e88dffe634030" gracePeriod=600 Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.516378 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="25c3576c8f6c0c703275007665c285ca572de78d4e00d849bb0e88dffe634030" exitCode=0 Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.516451 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"25c3576c8f6c0c703275007665c285ca572de78d4e00d849bb0e88dffe634030"} Nov 24 20:00:15 crc kubenswrapper[5035]: I1124 20:00:15.516730 5035 scope.go:117] "RemoveContainer" containerID="9432e7a82af869cd0cbfd07cf3972602c4f51ef2a914dcc9372e7765e3910c20" Nov 24 20:00:16 crc kubenswrapper[5035]: I1124 20:00:16.532549 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e"} Nov 24 20:00:27 crc kubenswrapper[5035]: I1124 20:00:27.430115 5035 scope.go:117] "RemoveContainer" containerID="29e626b8c17193b8af3af3bae25eabcf87a14bb8cd3890303cb7d224f46f468f" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.571933 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gtl5q"] Nov 24 20:00:33 crc kubenswrapper[5035]: E1124 20:00:33.573142 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72106bf6-5bff-4412-b58e-47c869df5338" containerName="collect-profiles" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.573163 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="72106bf6-5bff-4412-b58e-47c869df5338" containerName="collect-profiles" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.573538 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="72106bf6-5bff-4412-b58e-47c869df5338" containerName="collect-profiles" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.575917 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.584570 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtl5q"] Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.753110 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-catalog-content\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.753176 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grcjs\" (UniqueName: \"kubernetes.io/projected/1194c387-81c3-4f33-b29e-edbf90c3a49c-kube-api-access-grcjs\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.753209 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-utilities\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.854751 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-catalog-content\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.854818 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grcjs\" (UniqueName: \"kubernetes.io/projected/1194c387-81c3-4f33-b29e-edbf90c3a49c-kube-api-access-grcjs\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.854849 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-utilities\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.855369 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-utilities\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.855374 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-catalog-content\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.878160 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grcjs\" (UniqueName: \"kubernetes.io/projected/1194c387-81c3-4f33-b29e-edbf90c3a49c-kube-api-access-grcjs\") pod \"redhat-marketplace-gtl5q\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:33 crc kubenswrapper[5035]: I1124 20:00:33.960124 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:34 crc kubenswrapper[5035]: I1124 20:00:34.414034 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtl5q"] Nov 24 20:00:34 crc kubenswrapper[5035]: I1124 20:00:34.720925 5035 generic.go:334] "Generic (PLEG): container finished" podID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerID="70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0" exitCode=0 Nov 24 20:00:34 crc kubenswrapper[5035]: I1124 20:00:34.721039 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtl5q" event={"ID":"1194c387-81c3-4f33-b29e-edbf90c3a49c","Type":"ContainerDied","Data":"70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0"} Nov 24 20:00:34 crc kubenswrapper[5035]: I1124 20:00:34.721222 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtl5q" event={"ID":"1194c387-81c3-4f33-b29e-edbf90c3a49c","Type":"ContainerStarted","Data":"93fcd5044cc10d73bc0ce65e17a235121fdf1f62a6fb5552b4a6c16f0d4116e5"} Nov 24 20:00:34 crc kubenswrapper[5035]: I1124 20:00:34.724972 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:00:35 crc kubenswrapper[5035]: I1124 20:00:35.731891 5035 generic.go:334] "Generic (PLEG): container finished" podID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerID="b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd" exitCode=0 Nov 24 20:00:35 crc kubenswrapper[5035]: I1124 20:00:35.732147 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtl5q" event={"ID":"1194c387-81c3-4f33-b29e-edbf90c3a49c","Type":"ContainerDied","Data":"b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd"} Nov 24 20:00:36 crc kubenswrapper[5035]: I1124 20:00:36.743111 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtl5q" event={"ID":"1194c387-81c3-4f33-b29e-edbf90c3a49c","Type":"ContainerStarted","Data":"3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d"} Nov 24 20:00:36 crc kubenswrapper[5035]: I1124 20:00:36.796072 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gtl5q" podStartSLOduration=2.378399609 podStartE2EDuration="3.796052151s" podCreationTimestamp="2025-11-24 20:00:33 +0000 UTC" firstStartedPulling="2025-11-24 20:00:34.724704881 +0000 UTC m=+2773.247211138" lastFinishedPulling="2025-11-24 20:00:36.142357423 +0000 UTC m=+2774.664863680" observedRunningTime="2025-11-24 20:00:36.786913518 +0000 UTC m=+2775.309419775" watchObservedRunningTime="2025-11-24 20:00:36.796052151 +0000 UTC m=+2775.318558418" Nov 24 20:00:43 crc kubenswrapper[5035]: I1124 20:00:43.960524 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:43 crc kubenswrapper[5035]: I1124 20:00:43.961177 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:44 crc kubenswrapper[5035]: I1124 20:00:44.022097 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:44 crc kubenswrapper[5035]: I1124 20:00:44.877157 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:44 crc kubenswrapper[5035]: I1124 20:00:44.944649 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtl5q"] Nov 24 20:00:46 crc kubenswrapper[5035]: I1124 20:00:46.836520 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gtl5q" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="registry-server" containerID="cri-o://3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d" gracePeriod=2 Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.336667 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.415627 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grcjs\" (UniqueName: \"kubernetes.io/projected/1194c387-81c3-4f33-b29e-edbf90c3a49c-kube-api-access-grcjs\") pod \"1194c387-81c3-4f33-b29e-edbf90c3a49c\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.416014 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-catalog-content\") pod \"1194c387-81c3-4f33-b29e-edbf90c3a49c\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.416359 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-utilities\") pod \"1194c387-81c3-4f33-b29e-edbf90c3a49c\" (UID: \"1194c387-81c3-4f33-b29e-edbf90c3a49c\") " Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.417050 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-utilities" (OuterVolumeSpecName: "utilities") pod "1194c387-81c3-4f33-b29e-edbf90c3a49c" (UID: "1194c387-81c3-4f33-b29e-edbf90c3a49c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.417496 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.421585 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1194c387-81c3-4f33-b29e-edbf90c3a49c-kube-api-access-grcjs" (OuterVolumeSpecName: "kube-api-access-grcjs") pod "1194c387-81c3-4f33-b29e-edbf90c3a49c" (UID: "1194c387-81c3-4f33-b29e-edbf90c3a49c"). InnerVolumeSpecName "kube-api-access-grcjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.435259 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1194c387-81c3-4f33-b29e-edbf90c3a49c" (UID: "1194c387-81c3-4f33-b29e-edbf90c3a49c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.519003 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grcjs\" (UniqueName: \"kubernetes.io/projected/1194c387-81c3-4f33-b29e-edbf90c3a49c-kube-api-access-grcjs\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.519046 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1194c387-81c3-4f33-b29e-edbf90c3a49c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.851650 5035 generic.go:334] "Generic (PLEG): container finished" podID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerID="3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d" exitCode=0 Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.851710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtl5q" event={"ID":"1194c387-81c3-4f33-b29e-edbf90c3a49c","Type":"ContainerDied","Data":"3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d"} Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.851751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtl5q" event={"ID":"1194c387-81c3-4f33-b29e-edbf90c3a49c","Type":"ContainerDied","Data":"93fcd5044cc10d73bc0ce65e17a235121fdf1f62a6fb5552b4a6c16f0d4116e5"} Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.851783 5035 scope.go:117] "RemoveContainer" containerID="3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.851815 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtl5q" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.887379 5035 scope.go:117] "RemoveContainer" containerID="b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.912908 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtl5q"] Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.919422 5035 scope.go:117] "RemoveContainer" containerID="70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.925251 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtl5q"] Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.963336 5035 scope.go:117] "RemoveContainer" containerID="3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d" Nov 24 20:00:47 crc kubenswrapper[5035]: E1124 20:00:47.963835 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d\": container with ID starting with 3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d not found: ID does not exist" containerID="3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.963874 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d"} err="failed to get container status \"3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d\": rpc error: code = NotFound desc = could not find container \"3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d\": container with ID starting with 3ec0bc0bc6053db2423fef92a31df7b4d5a6b871a81b0d2f382569ae06ad1a2d not found: ID does not exist" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.963900 5035 scope.go:117] "RemoveContainer" containerID="b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd" Nov 24 20:00:47 crc kubenswrapper[5035]: E1124 20:00:47.964187 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd\": container with ID starting with b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd not found: ID does not exist" containerID="b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.964226 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd"} err="failed to get container status \"b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd\": rpc error: code = NotFound desc = could not find container \"b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd\": container with ID starting with b7cf56b2904b9cae4e471bd861df9d89d03e409f48200a01c4f179f75064d9cd not found: ID does not exist" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.964253 5035 scope.go:117] "RemoveContainer" containerID="70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0" Nov 24 20:00:47 crc kubenswrapper[5035]: E1124 20:00:47.964516 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0\": container with ID starting with 70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0 not found: ID does not exist" containerID="70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0" Nov 24 20:00:47 crc kubenswrapper[5035]: I1124 20:00:47.964538 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0"} err="failed to get container status \"70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0\": rpc error: code = NotFound desc = could not find container \"70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0\": container with ID starting with 70255647b6f3b6c210862f6ae64d3a967792cae85e41d832337c5afa198ed9a0 not found: ID does not exist" Nov 24 20:00:48 crc kubenswrapper[5035]: I1124 20:00:48.223788 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" path="/var/lib/kubelet/pods/1194c387-81c3-4f33-b29e-edbf90c3a49c/volumes" Nov 24 20:00:56 crc kubenswrapper[5035]: I1124 20:00:56.934865 5035 generic.go:334] "Generic (PLEG): container finished" podID="f2273d9c-4343-4bb5-9bca-0a87c601cf15" containerID="9da177310c63d9d083b4288395c6548859d31059e826eac6335e170c5d099930" exitCode=0 Nov 24 20:00:56 crc kubenswrapper[5035]: I1124 20:00:56.934986 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" event={"ID":"f2273d9c-4343-4bb5-9bca-0a87c601cf15","Type":"ContainerDied","Data":"9da177310c63d9d083b4288395c6548859d31059e826eac6335e170c5d099930"} Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.354779 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.455480 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-inventory\") pod \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.455541 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovncontroller-config-0\") pod \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.455600 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pktlh\" (UniqueName: \"kubernetes.io/projected/f2273d9c-4343-4bb5-9bca-0a87c601cf15-kube-api-access-pktlh\") pod \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.455687 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovn-combined-ca-bundle\") pod \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.455716 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ssh-key\") pod \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.455830 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ceph\") pod \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\" (UID: \"f2273d9c-4343-4bb5-9bca-0a87c601cf15\") " Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.467496 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ceph" (OuterVolumeSpecName: "ceph") pod "f2273d9c-4343-4bb5-9bca-0a87c601cf15" (UID: "f2273d9c-4343-4bb5-9bca-0a87c601cf15"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.467627 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f2273d9c-4343-4bb5-9bca-0a87c601cf15" (UID: "f2273d9c-4343-4bb5-9bca-0a87c601cf15"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.467627 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2273d9c-4343-4bb5-9bca-0a87c601cf15-kube-api-access-pktlh" (OuterVolumeSpecName: "kube-api-access-pktlh") pod "f2273d9c-4343-4bb5-9bca-0a87c601cf15" (UID: "f2273d9c-4343-4bb5-9bca-0a87c601cf15"). InnerVolumeSpecName "kube-api-access-pktlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.481031 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "f2273d9c-4343-4bb5-9bca-0a87c601cf15" (UID: "f2273d9c-4343-4bb5-9bca-0a87c601cf15"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.483334 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f2273d9c-4343-4bb5-9bca-0a87c601cf15" (UID: "f2273d9c-4343-4bb5-9bca-0a87c601cf15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.484497 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-inventory" (OuterVolumeSpecName: "inventory") pod "f2273d9c-4343-4bb5-9bca-0a87c601cf15" (UID: "f2273d9c-4343-4bb5-9bca-0a87c601cf15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.558365 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.558394 5035 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.558404 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pktlh\" (UniqueName: \"kubernetes.io/projected/f2273d9c-4343-4bb5-9bca-0a87c601cf15-kube-api-access-pktlh\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.558413 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.558423 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.558430 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f2273d9c-4343-4bb5-9bca-0a87c601cf15-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.954360 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" event={"ID":"f2273d9c-4343-4bb5-9bca-0a87c601cf15","Type":"ContainerDied","Data":"ff3b88161d157c558fd88c4832632cbf9c68295e28eeac6b56f1c6f907e41558"} Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.954793 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff3b88161d157c558fd88c4832632cbf9c68295e28eeac6b56f1c6f907e41558" Nov 24 20:00:58 crc kubenswrapper[5035]: I1124 20:00:58.954398 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-b4rlz" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.087846 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48"] Nov 24 20:00:59 crc kubenswrapper[5035]: E1124 20:00:59.088211 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="registry-server" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.088227 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="registry-server" Nov 24 20:00:59 crc kubenswrapper[5035]: E1124 20:00:59.088263 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="extract-content" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.088269 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="extract-content" Nov 24 20:00:59 crc kubenswrapper[5035]: E1124 20:00:59.088280 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="extract-utilities" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.088286 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="extract-utilities" Nov 24 20:00:59 crc kubenswrapper[5035]: E1124 20:00:59.088297 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2273d9c-4343-4bb5-9bca-0a87c601cf15" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.088303 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2273d9c-4343-4bb5-9bca-0a87c601cf15" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.088478 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1194c387-81c3-4f33-b29e-edbf90c3a49c" containerName="registry-server" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.088500 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2273d9c-4343-4bb5-9bca-0a87c601cf15" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.089103 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.094524 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.094609 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.094549 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.094718 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.094958 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.095197 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.095542 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.098725 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48"] Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.169977 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.170260 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.170450 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.170549 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.170729 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.170765 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.170831 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz5sf\" (UniqueName: \"kubernetes.io/projected/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-kube-api-access-lz5sf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.271817 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.271958 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.272007 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.272036 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.272222 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.272264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz5sf\" (UniqueName: \"kubernetes.io/projected/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-kube-api-access-lz5sf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.272291 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.278042 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.278716 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.278785 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.279029 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.279097 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.281002 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.291899 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz5sf\" (UniqueName: \"kubernetes.io/projected/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-kube-api-access-lz5sf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.410817 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.941856 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48"] Nov 24 20:00:59 crc kubenswrapper[5035]: I1124 20:00:59.966381 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" event={"ID":"ecfdf321-e4fd-4703-a5cd-907d6508d7e6","Type":"ContainerStarted","Data":"07c7d423f57abbdf401ecb637911dec978f1b087f6e3bc954ff748bc77972f44"} Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.125232 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29400241-gqccc"] Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.126278 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.145185 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400241-gqccc"] Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.192919 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-fernet-keys\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.193086 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-combined-ca-bundle\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.193388 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-config-data\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.193460 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz6l5\" (UniqueName: \"kubernetes.io/projected/549314e7-60d0-4aa6-a48b-a36acdaefa29-kube-api-access-kz6l5\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.295679 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-config-data\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.295725 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz6l5\" (UniqueName: \"kubernetes.io/projected/549314e7-60d0-4aa6-a48b-a36acdaefa29-kube-api-access-kz6l5\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.295773 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-fernet-keys\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.295872 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-combined-ca-bundle\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.301739 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-combined-ca-bundle\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.301786 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-fernet-keys\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.302120 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-config-data\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.311248 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz6l5\" (UniqueName: \"kubernetes.io/projected/549314e7-60d0-4aa6-a48b-a36acdaefa29-kube-api-access-kz6l5\") pod \"keystone-cron-29400241-gqccc\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.499401 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:00 crc kubenswrapper[5035]: W1124 20:01:00.926386 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod549314e7_60d0_4aa6_a48b_a36acdaefa29.slice/crio-32ed904e97fd03df2c5fd82de215112d7ef78de41dcafa222e3dce04dd4976c4 WatchSource:0}: Error finding container 32ed904e97fd03df2c5fd82de215112d7ef78de41dcafa222e3dce04dd4976c4: Status 404 returned error can't find the container with id 32ed904e97fd03df2c5fd82de215112d7ef78de41dcafa222e3dce04dd4976c4 Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.926496 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400241-gqccc"] Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.975709 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400241-gqccc" event={"ID":"549314e7-60d0-4aa6-a48b-a36acdaefa29","Type":"ContainerStarted","Data":"32ed904e97fd03df2c5fd82de215112d7ef78de41dcafa222e3dce04dd4976c4"} Nov 24 20:01:00 crc kubenswrapper[5035]: I1124 20:01:00.977644 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" event={"ID":"ecfdf321-e4fd-4703-a5cd-907d6508d7e6","Type":"ContainerStarted","Data":"5d8365c14ce38688cc1f1205ddbedc14220175d66d8fe278572a136b26fc6482"} Nov 24 20:01:01 crc kubenswrapper[5035]: I1124 20:01:01.000049 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" podStartSLOduration=1.501728768 podStartE2EDuration="2.000024958s" podCreationTimestamp="2025-11-24 20:00:59 +0000 UTC" firstStartedPulling="2025-11-24 20:00:59.952533433 +0000 UTC m=+2798.475039690" lastFinishedPulling="2025-11-24 20:01:00.450829613 +0000 UTC m=+2798.973335880" observedRunningTime="2025-11-24 20:01:01.000019978 +0000 UTC m=+2799.522526295" watchObservedRunningTime="2025-11-24 20:01:01.000024958 +0000 UTC m=+2799.522531255" Nov 24 20:01:02 crc kubenswrapper[5035]: I1124 20:01:02.001687 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400241-gqccc" event={"ID":"549314e7-60d0-4aa6-a48b-a36acdaefa29","Type":"ContainerStarted","Data":"d6c3d054444c58edd092415d46a654c9b973db79a03dcf2cee736c6936f12106"} Nov 24 20:01:02 crc kubenswrapper[5035]: I1124 20:01:02.026758 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29400241-gqccc" podStartSLOduration=2.026734309 podStartE2EDuration="2.026734309s" podCreationTimestamp="2025-11-24 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:01:02.019867176 +0000 UTC m=+2800.542373483" watchObservedRunningTime="2025-11-24 20:01:02.026734309 +0000 UTC m=+2800.549240566" Nov 24 20:01:04 crc kubenswrapper[5035]: I1124 20:01:04.021263 5035 generic.go:334] "Generic (PLEG): container finished" podID="549314e7-60d0-4aa6-a48b-a36acdaefa29" containerID="d6c3d054444c58edd092415d46a654c9b973db79a03dcf2cee736c6936f12106" exitCode=0 Nov 24 20:01:04 crc kubenswrapper[5035]: I1124 20:01:04.021389 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400241-gqccc" event={"ID":"549314e7-60d0-4aa6-a48b-a36acdaefa29","Type":"ContainerDied","Data":"d6c3d054444c58edd092415d46a654c9b973db79a03dcf2cee736c6936f12106"} Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.344937 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.494651 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz6l5\" (UniqueName: \"kubernetes.io/projected/549314e7-60d0-4aa6-a48b-a36acdaefa29-kube-api-access-kz6l5\") pod \"549314e7-60d0-4aa6-a48b-a36acdaefa29\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.494957 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-combined-ca-bundle\") pod \"549314e7-60d0-4aa6-a48b-a36acdaefa29\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.495064 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-config-data\") pod \"549314e7-60d0-4aa6-a48b-a36acdaefa29\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.495219 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-fernet-keys\") pod \"549314e7-60d0-4aa6-a48b-a36acdaefa29\" (UID: \"549314e7-60d0-4aa6-a48b-a36acdaefa29\") " Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.501402 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "549314e7-60d0-4aa6-a48b-a36acdaefa29" (UID: "549314e7-60d0-4aa6-a48b-a36acdaefa29"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.502542 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549314e7-60d0-4aa6-a48b-a36acdaefa29-kube-api-access-kz6l5" (OuterVolumeSpecName: "kube-api-access-kz6l5") pod "549314e7-60d0-4aa6-a48b-a36acdaefa29" (UID: "549314e7-60d0-4aa6-a48b-a36acdaefa29"). InnerVolumeSpecName "kube-api-access-kz6l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.551421 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "549314e7-60d0-4aa6-a48b-a36acdaefa29" (UID: "549314e7-60d0-4aa6-a48b-a36acdaefa29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.576345 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-config-data" (OuterVolumeSpecName: "config-data") pod "549314e7-60d0-4aa6-a48b-a36acdaefa29" (UID: "549314e7-60d0-4aa6-a48b-a36acdaefa29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.597533 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz6l5\" (UniqueName: \"kubernetes.io/projected/549314e7-60d0-4aa6-a48b-a36acdaefa29-kube-api-access-kz6l5\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.597595 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.597620 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:05 crc kubenswrapper[5035]: I1124 20:01:05.597645 5035 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/549314e7-60d0-4aa6-a48b-a36acdaefa29-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:06 crc kubenswrapper[5035]: I1124 20:01:06.042551 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400241-gqccc" event={"ID":"549314e7-60d0-4aa6-a48b-a36acdaefa29","Type":"ContainerDied","Data":"32ed904e97fd03df2c5fd82de215112d7ef78de41dcafa222e3dce04dd4976c4"} Nov 24 20:01:06 crc kubenswrapper[5035]: I1124 20:01:06.042824 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32ed904e97fd03df2c5fd82de215112d7ef78de41dcafa222e3dce04dd4976c4" Nov 24 20:01:06 crc kubenswrapper[5035]: I1124 20:01:06.042897 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400241-gqccc" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.235707 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f5g2p"] Nov 24 20:01:31 crc kubenswrapper[5035]: E1124 20:01:31.236772 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549314e7-60d0-4aa6-a48b-a36acdaefa29" containerName="keystone-cron" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.236802 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="549314e7-60d0-4aa6-a48b-a36acdaefa29" containerName="keystone-cron" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.237084 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="549314e7-60d0-4aa6-a48b-a36acdaefa29" containerName="keystone-cron" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.238896 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.259826 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5g2p"] Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.387659 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-utilities\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.387868 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpfc2\" (UniqueName: \"kubernetes.io/projected/133ce13f-8dde-419d-86b9-61c69bee387e-kube-api-access-qpfc2\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.387948 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-catalog-content\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.490034 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpfc2\" (UniqueName: \"kubernetes.io/projected/133ce13f-8dde-419d-86b9-61c69bee387e-kube-api-access-qpfc2\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.490117 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-catalog-content\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.490173 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-utilities\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.490698 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-catalog-content\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.490738 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-utilities\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.512842 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpfc2\" (UniqueName: \"kubernetes.io/projected/133ce13f-8dde-419d-86b9-61c69bee387e-kube-api-access-qpfc2\") pod \"community-operators-f5g2p\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.558039 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:31 crc kubenswrapper[5035]: I1124 20:01:31.862853 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5g2p"] Nov 24 20:01:32 crc kubenswrapper[5035]: I1124 20:01:32.297618 5035 generic.go:334] "Generic (PLEG): container finished" podID="133ce13f-8dde-419d-86b9-61c69bee387e" containerID="65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda" exitCode=0 Nov 24 20:01:32 crc kubenswrapper[5035]: I1124 20:01:32.297676 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5g2p" event={"ID":"133ce13f-8dde-419d-86b9-61c69bee387e","Type":"ContainerDied","Data":"65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda"} Nov 24 20:01:32 crc kubenswrapper[5035]: I1124 20:01:32.297711 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5g2p" event={"ID":"133ce13f-8dde-419d-86b9-61c69bee387e","Type":"ContainerStarted","Data":"6b3025c647bab9571a2cfde78eb6abddd8eba169cf9a45e364e6b65e4ef85ef8"} Nov 24 20:01:34 crc kubenswrapper[5035]: I1124 20:01:34.318061 5035 generic.go:334] "Generic (PLEG): container finished" podID="133ce13f-8dde-419d-86b9-61c69bee387e" containerID="379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5" exitCode=0 Nov 24 20:01:34 crc kubenswrapper[5035]: I1124 20:01:34.318164 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5g2p" event={"ID":"133ce13f-8dde-419d-86b9-61c69bee387e","Type":"ContainerDied","Data":"379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5"} Nov 24 20:01:35 crc kubenswrapper[5035]: I1124 20:01:35.332210 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5g2p" event={"ID":"133ce13f-8dde-419d-86b9-61c69bee387e","Type":"ContainerStarted","Data":"4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc"} Nov 24 20:01:35 crc kubenswrapper[5035]: I1124 20:01:35.360875 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f5g2p" podStartSLOduration=1.835408583 podStartE2EDuration="4.360857257s" podCreationTimestamp="2025-11-24 20:01:31 +0000 UTC" firstStartedPulling="2025-11-24 20:01:32.299230344 +0000 UTC m=+2830.821736601" lastFinishedPulling="2025-11-24 20:01:34.824679018 +0000 UTC m=+2833.347185275" observedRunningTime="2025-11-24 20:01:35.357743544 +0000 UTC m=+2833.880249841" watchObservedRunningTime="2025-11-24 20:01:35.360857257 +0000 UTC m=+2833.883363514" Nov 24 20:01:37 crc kubenswrapper[5035]: I1124 20:01:37.838423 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-68r24"] Nov 24 20:01:37 crc kubenswrapper[5035]: I1124 20:01:37.843351 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:37 crc kubenswrapper[5035]: I1124 20:01:37.847188 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68r24"] Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.027879 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n85h\" (UniqueName: \"kubernetes.io/projected/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-kube-api-access-7n85h\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.027947 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-catalog-content\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.028462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-utilities\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.129575 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n85h\" (UniqueName: \"kubernetes.io/projected/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-kube-api-access-7n85h\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.129640 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-catalog-content\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.129668 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-utilities\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.130225 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-catalog-content\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.130273 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-utilities\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.159940 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n85h\" (UniqueName: \"kubernetes.io/projected/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-kube-api-access-7n85h\") pod \"certified-operators-68r24\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.172394 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:38 crc kubenswrapper[5035]: I1124 20:01:38.652857 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68r24"] Nov 24 20:01:38 crc kubenswrapper[5035]: W1124 20:01:38.656660 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d58e2eb_879c_4316_9bd9_f0f7ed07fcfa.slice/crio-5f1c234536ab5636dc42dffc1d63bbd0a6d9cb0c422c47065b38d1e0f1dde09d WatchSource:0}: Error finding container 5f1c234536ab5636dc42dffc1d63bbd0a6d9cb0c422c47065b38d1e0f1dde09d: Status 404 returned error can't find the container with id 5f1c234536ab5636dc42dffc1d63bbd0a6d9cb0c422c47065b38d1e0f1dde09d Nov 24 20:01:39 crc kubenswrapper[5035]: I1124 20:01:39.574752 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerID="1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8" exitCode=0 Nov 24 20:01:39 crc kubenswrapper[5035]: I1124 20:01:39.574826 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68r24" event={"ID":"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa","Type":"ContainerDied","Data":"1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8"} Nov 24 20:01:39 crc kubenswrapper[5035]: I1124 20:01:39.575105 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68r24" event={"ID":"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa","Type":"ContainerStarted","Data":"5f1c234536ab5636dc42dffc1d63bbd0a6d9cb0c422c47065b38d1e0f1dde09d"} Nov 24 20:01:41 crc kubenswrapper[5035]: I1124 20:01:41.558383 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:41 crc kubenswrapper[5035]: I1124 20:01:41.558658 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:41 crc kubenswrapper[5035]: I1124 20:01:41.602084 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerID="d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766" exitCode=0 Nov 24 20:01:41 crc kubenswrapper[5035]: I1124 20:01:41.602440 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68r24" event={"ID":"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa","Type":"ContainerDied","Data":"d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766"} Nov 24 20:01:41 crc kubenswrapper[5035]: I1124 20:01:41.618158 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:41 crc kubenswrapper[5035]: I1124 20:01:41.673716 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:42 crc kubenswrapper[5035]: I1124 20:01:42.616440 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68r24" event={"ID":"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa","Type":"ContainerStarted","Data":"ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd"} Nov 24 20:01:42 crc kubenswrapper[5035]: I1124 20:01:42.639937 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-68r24" podStartSLOduration=3.231847793 podStartE2EDuration="5.63991206s" podCreationTimestamp="2025-11-24 20:01:37 +0000 UTC" firstStartedPulling="2025-11-24 20:01:39.57837068 +0000 UTC m=+2838.100876977" lastFinishedPulling="2025-11-24 20:01:41.986434987 +0000 UTC m=+2840.508941244" observedRunningTime="2025-11-24 20:01:42.63315825 +0000 UTC m=+2841.155664517" watchObservedRunningTime="2025-11-24 20:01:42.63991206 +0000 UTC m=+2841.162418327" Nov 24 20:01:43 crc kubenswrapper[5035]: I1124 20:01:43.611339 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5g2p"] Nov 24 20:01:43 crc kubenswrapper[5035]: I1124 20:01:43.624334 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f5g2p" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="registry-server" containerID="cri-o://4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc" gracePeriod=2 Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.085521 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.155026 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-utilities\") pod \"133ce13f-8dde-419d-86b9-61c69bee387e\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.155793 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-utilities" (OuterVolumeSpecName: "utilities") pod "133ce13f-8dde-419d-86b9-61c69bee387e" (UID: "133ce13f-8dde-419d-86b9-61c69bee387e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.155989 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpfc2\" (UniqueName: \"kubernetes.io/projected/133ce13f-8dde-419d-86b9-61c69bee387e-kube-api-access-qpfc2\") pod \"133ce13f-8dde-419d-86b9-61c69bee387e\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.156738 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-catalog-content\") pod \"133ce13f-8dde-419d-86b9-61c69bee387e\" (UID: \"133ce13f-8dde-419d-86b9-61c69bee387e\") " Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.162330 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/133ce13f-8dde-419d-86b9-61c69bee387e-kube-api-access-qpfc2" (OuterVolumeSpecName: "kube-api-access-qpfc2") pod "133ce13f-8dde-419d-86b9-61c69bee387e" (UID: "133ce13f-8dde-419d-86b9-61c69bee387e"). InnerVolumeSpecName "kube-api-access-qpfc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.164921 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpfc2\" (UniqueName: \"kubernetes.io/projected/133ce13f-8dde-419d-86b9-61c69bee387e-kube-api-access-qpfc2\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.164968 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.207595 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "133ce13f-8dde-419d-86b9-61c69bee387e" (UID: "133ce13f-8dde-419d-86b9-61c69bee387e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.266697 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/133ce13f-8dde-419d-86b9-61c69bee387e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.638996 5035 generic.go:334] "Generic (PLEG): container finished" podID="133ce13f-8dde-419d-86b9-61c69bee387e" containerID="4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc" exitCode=0 Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.639067 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5g2p" event={"ID":"133ce13f-8dde-419d-86b9-61c69bee387e","Type":"ContainerDied","Data":"4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc"} Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.639093 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5g2p" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.640777 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5g2p" event={"ID":"133ce13f-8dde-419d-86b9-61c69bee387e","Type":"ContainerDied","Data":"6b3025c647bab9571a2cfde78eb6abddd8eba169cf9a45e364e6b65e4ef85ef8"} Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.640796 5035 scope.go:117] "RemoveContainer" containerID="4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.674762 5035 scope.go:117] "RemoveContainer" containerID="379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.674816 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5g2p"] Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.687317 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f5g2p"] Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.702798 5035 scope.go:117] "RemoveContainer" containerID="65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.748792 5035 scope.go:117] "RemoveContainer" containerID="4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc" Nov 24 20:01:44 crc kubenswrapper[5035]: E1124 20:01:44.749163 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc\": container with ID starting with 4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc not found: ID does not exist" containerID="4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.749203 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc"} err="failed to get container status \"4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc\": rpc error: code = NotFound desc = could not find container \"4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc\": container with ID starting with 4271f7dccd91e80e7e25ba11dccca44636de523c82097bc86d3a1f375da018bc not found: ID does not exist" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.749227 5035 scope.go:117] "RemoveContainer" containerID="379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5" Nov 24 20:01:44 crc kubenswrapper[5035]: E1124 20:01:44.749507 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5\": container with ID starting with 379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5 not found: ID does not exist" containerID="379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.749545 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5"} err="failed to get container status \"379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5\": rpc error: code = NotFound desc = could not find container \"379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5\": container with ID starting with 379fc0b3a527e4f9478d73e8389de7bc9567bf106e716ef810c7c254604755f5 not found: ID does not exist" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.749564 5035 scope.go:117] "RemoveContainer" containerID="65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda" Nov 24 20:01:44 crc kubenswrapper[5035]: E1124 20:01:44.749919 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda\": container with ID starting with 65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda not found: ID does not exist" containerID="65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda" Nov 24 20:01:44 crc kubenswrapper[5035]: I1124 20:01:44.749957 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda"} err="failed to get container status \"65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda\": rpc error: code = NotFound desc = could not find container \"65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda\": container with ID starting with 65798dfbfbf78fbb10867e751eeffdd7cdf90a987cd6558ddaab3cb2c3decbda not found: ID does not exist" Nov 24 20:01:46 crc kubenswrapper[5035]: I1124 20:01:46.213914 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" path="/var/lib/kubelet/pods/133ce13f-8dde-419d-86b9-61c69bee387e/volumes" Nov 24 20:01:48 crc kubenswrapper[5035]: I1124 20:01:48.172626 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:48 crc kubenswrapper[5035]: I1124 20:01:48.172678 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:48 crc kubenswrapper[5035]: I1124 20:01:48.234996 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:48 crc kubenswrapper[5035]: I1124 20:01:48.756697 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:48 crc kubenswrapper[5035]: I1124 20:01:48.801353 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68r24"] Nov 24 20:01:50 crc kubenswrapper[5035]: I1124 20:01:50.703496 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-68r24" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="registry-server" containerID="cri-o://ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd" gracePeriod=2 Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.279669 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.402091 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-catalog-content\") pod \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.402204 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-utilities\") pod \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.402441 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n85h\" (UniqueName: \"kubernetes.io/projected/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-kube-api-access-7n85h\") pod \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\" (UID: \"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa\") " Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.403250 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-utilities" (OuterVolumeSpecName: "utilities") pod "4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" (UID: "4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.408312 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-kube-api-access-7n85h" (OuterVolumeSpecName: "kube-api-access-7n85h") pod "4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" (UID: "4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa"). InnerVolumeSpecName "kube-api-access-7n85h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.453140 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" (UID: "4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.504696 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n85h\" (UniqueName: \"kubernetes.io/projected/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-kube-api-access-7n85h\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.504730 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.504741 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.714020 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerID="ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd" exitCode=0 Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.714059 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68r24" event={"ID":"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa","Type":"ContainerDied","Data":"ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd"} Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.714100 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68r24" event={"ID":"4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa","Type":"ContainerDied","Data":"5f1c234536ab5636dc42dffc1d63bbd0a6d9cb0c422c47065b38d1e0f1dde09d"} Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.714117 5035 scope.go:117] "RemoveContainer" containerID="ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.714265 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68r24" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.749350 5035 scope.go:117] "RemoveContainer" containerID="d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.761937 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68r24"] Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.768986 5035 scope.go:117] "RemoveContainer" containerID="1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.777080 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-68r24"] Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.813662 5035 scope.go:117] "RemoveContainer" containerID="ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd" Nov 24 20:01:51 crc kubenswrapper[5035]: E1124 20:01:51.814448 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd\": container with ID starting with ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd not found: ID does not exist" containerID="ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.814489 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd"} err="failed to get container status \"ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd\": rpc error: code = NotFound desc = could not find container \"ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd\": container with ID starting with ff038b2cfeda0b0d4a0d834ac1d9689115442eb4231892e2d3769f887c2fc4cd not found: ID does not exist" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.814516 5035 scope.go:117] "RemoveContainer" containerID="d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766" Nov 24 20:01:51 crc kubenswrapper[5035]: E1124 20:01:51.814944 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766\": container with ID starting with d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766 not found: ID does not exist" containerID="d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.815060 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766"} err="failed to get container status \"d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766\": rpc error: code = NotFound desc = could not find container \"d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766\": container with ID starting with d65c7af7bbc066e4ae7e98202035832bac279ab78586cd863707cc792a5a7766 not found: ID does not exist" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.815165 5035 scope.go:117] "RemoveContainer" containerID="1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8" Nov 24 20:01:51 crc kubenswrapper[5035]: E1124 20:01:51.815871 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8\": container with ID starting with 1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8 not found: ID does not exist" containerID="1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8" Nov 24 20:01:51 crc kubenswrapper[5035]: I1124 20:01:51.815897 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8"} err="failed to get container status \"1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8\": rpc error: code = NotFound desc = could not find container \"1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8\": container with ID starting with 1a72e198a188071baeb507a671a36bfe16833a1e59e4806310186ecad59cdbb8 not found: ID does not exist" Nov 24 20:01:52 crc kubenswrapper[5035]: I1124 20:01:52.214341 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" path="/var/lib/kubelet/pods/4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa/volumes" Nov 24 20:02:08 crc kubenswrapper[5035]: I1124 20:02:08.891837 5035 generic.go:334] "Generic (PLEG): container finished" podID="ecfdf321-e4fd-4703-a5cd-907d6508d7e6" containerID="5d8365c14ce38688cc1f1205ddbedc14220175d66d8fe278572a136b26fc6482" exitCode=0 Nov 24 20:02:08 crc kubenswrapper[5035]: I1124 20:02:08.891947 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" event={"ID":"ecfdf321-e4fd-4703-a5cd-907d6508d7e6","Type":"ContainerDied","Data":"5d8365c14ce38688cc1f1205ddbedc14220175d66d8fe278572a136b26fc6482"} Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.386572 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.527901 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-metadata-combined-ca-bundle\") pod \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.528433 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ceph\") pod \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.528720 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-inventory\") pod \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.528992 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-nova-metadata-neutron-config-0\") pod \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.529275 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ssh-key\") pod \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.529579 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.529858 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz5sf\" (UniqueName: \"kubernetes.io/projected/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-kube-api-access-lz5sf\") pod \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\" (UID: \"ecfdf321-e4fd-4703-a5cd-907d6508d7e6\") " Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.534235 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-kube-api-access-lz5sf" (OuterVolumeSpecName: "kube-api-access-lz5sf") pod "ecfdf321-e4fd-4703-a5cd-907d6508d7e6" (UID: "ecfdf321-e4fd-4703-a5cd-907d6508d7e6"). InnerVolumeSpecName "kube-api-access-lz5sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.535313 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ecfdf321-e4fd-4703-a5cd-907d6508d7e6" (UID: "ecfdf321-e4fd-4703-a5cd-907d6508d7e6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.536159 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ceph" (OuterVolumeSpecName: "ceph") pod "ecfdf321-e4fd-4703-a5cd-907d6508d7e6" (UID: "ecfdf321-e4fd-4703-a5cd-907d6508d7e6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.565130 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ecfdf321-e4fd-4703-a5cd-907d6508d7e6" (UID: "ecfdf321-e4fd-4703-a5cd-907d6508d7e6"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.579349 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ecfdf321-e4fd-4703-a5cd-907d6508d7e6" (UID: "ecfdf321-e4fd-4703-a5cd-907d6508d7e6"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.586588 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-inventory" (OuterVolumeSpecName: "inventory") pod "ecfdf321-e4fd-4703-a5cd-907d6508d7e6" (UID: "ecfdf321-e4fd-4703-a5cd-907d6508d7e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.588607 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ecfdf321-e4fd-4703-a5cd-907d6508d7e6" (UID: "ecfdf321-e4fd-4703-a5cd-907d6508d7e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.632982 5035 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.633016 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.633032 5035 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.633046 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz5sf\" (UniqueName: \"kubernetes.io/projected/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-kube-api-access-lz5sf\") on node \"crc\" DevicePath \"\"" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.633059 5035 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.633071 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.633081 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecfdf321-e4fd-4703-a5cd-907d6508d7e6-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.922740 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" event={"ID":"ecfdf321-e4fd-4703-a5cd-907d6508d7e6","Type":"ContainerDied","Data":"07c7d423f57abbdf401ecb637911dec978f1b087f6e3bc954ff748bc77972f44"} Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.922804 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07c7d423f57abbdf401ecb637911dec978f1b087f6e3bc954ff748bc77972f44" Nov 24 20:02:10 crc kubenswrapper[5035]: I1124 20:02:10.922940 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.036169 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957"] Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.040172 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="extract-utilities" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040210 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="extract-utilities" Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.040230 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="registry-server" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040239 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="registry-server" Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.040265 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="registry-server" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040273 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="registry-server" Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.040322 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="extract-content" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040331 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="extract-content" Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.040342 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="extract-content" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040350 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="extract-content" Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.040362 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfdf321-e4fd-4703-a5cd-907d6508d7e6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040373 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfdf321-e4fd-4703-a5cd-907d6508d7e6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.040387 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="extract-utilities" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040395 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="extract-utilities" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040616 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d58e2eb-879c-4316-9bd9-f0f7ed07fcfa" containerName="registry-server" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040628 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfdf321-e4fd-4703-a5cd-907d6508d7e6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.040640 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="133ce13f-8dde-419d-86b9-61c69bee387e" containerName="registry-server" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.041458 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.049474 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.053124 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.053249 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.053435 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.053490 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.054859 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.088559 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957"] Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.144789 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.144924 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.144968 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7lhf\" (UniqueName: \"kubernetes.io/projected/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-kube-api-access-l7lhf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.144993 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.145029 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.145081 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: E1124 20:02:11.158319 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecfdf321_e4fd_4703_a5cd_907d6508d7e6.slice\": RecentStats: unable to find data in memory cache]" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.247116 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.247188 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7lhf\" (UniqueName: \"kubernetes.io/projected/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-kube-api-access-l7lhf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.247223 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.247276 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.247399 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.247451 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.252117 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.252160 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.252579 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.255281 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.257558 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.274915 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7lhf\" (UniqueName: \"kubernetes.io/projected/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-kube-api-access-l7lhf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fj957\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.379783 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.924421 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957"] Nov 24 20:02:11 crc kubenswrapper[5035]: I1124 20:02:11.940612 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" event={"ID":"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d","Type":"ContainerStarted","Data":"3770bb5a4177d426ebee32563d95b6723b89f0a94cdf0f65bdc5a16c334bd62d"} Nov 24 20:02:12 crc kubenswrapper[5035]: I1124 20:02:12.952003 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" event={"ID":"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d","Type":"ContainerStarted","Data":"6841bb91fdfa12418a914ec6ab45aff1074be057817590543f18798be685acef"} Nov 24 20:02:13 crc kubenswrapper[5035]: I1124 20:02:12.999955 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" podStartSLOduration=1.544884937 podStartE2EDuration="1.999920795s" podCreationTimestamp="2025-11-24 20:02:11 +0000 UTC" firstStartedPulling="2025-11-24 20:02:11.930054264 +0000 UTC m=+2870.452560521" lastFinishedPulling="2025-11-24 20:02:12.385090122 +0000 UTC m=+2870.907596379" observedRunningTime="2025-11-24 20:02:12.978907725 +0000 UTC m=+2871.501414022" watchObservedRunningTime="2025-11-24 20:02:12.999920795 +0000 UTC m=+2871.522427092" Nov 24 20:02:15 crc kubenswrapper[5035]: I1124 20:02:15.234506 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:02:15 crc kubenswrapper[5035]: I1124 20:02:15.235472 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:02:45 crc kubenswrapper[5035]: I1124 20:02:45.233962 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:02:45 crc kubenswrapper[5035]: I1124 20:02:45.234494 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.234723 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.235262 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.235320 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.235961 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.236017 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" gracePeriod=600 Nov 24 20:03:15 crc kubenswrapper[5035]: E1124 20:03:15.357514 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.609703 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" exitCode=0 Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.609779 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e"} Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.610062 5035 scope.go:117] "RemoveContainer" containerID="25c3576c8f6c0c703275007665c285ca572de78d4e00d849bb0e88dffe634030" Nov 24 20:03:15 crc kubenswrapper[5035]: I1124 20:03:15.611309 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:03:15 crc kubenswrapper[5035]: E1124 20:03:15.611610 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:03:29 crc kubenswrapper[5035]: I1124 20:03:29.200708 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:03:29 crc kubenswrapper[5035]: E1124 20:03:29.201539 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:03:41 crc kubenswrapper[5035]: I1124 20:03:41.200650 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:03:41 crc kubenswrapper[5035]: E1124 20:03:41.201426 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:03:54 crc kubenswrapper[5035]: I1124 20:03:54.201750 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:03:54 crc kubenswrapper[5035]: E1124 20:03:54.203012 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:04:06 crc kubenswrapper[5035]: I1124 20:04:06.200808 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:04:06 crc kubenswrapper[5035]: E1124 20:04:06.202413 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:04:19 crc kubenswrapper[5035]: I1124 20:04:19.200688 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:04:19 crc kubenswrapper[5035]: E1124 20:04:19.201393 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:04:33 crc kubenswrapper[5035]: I1124 20:04:33.200101 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:04:33 crc kubenswrapper[5035]: E1124 20:04:33.201113 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:04:45 crc kubenswrapper[5035]: I1124 20:04:45.199905 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:04:45 crc kubenswrapper[5035]: E1124 20:04:45.200797 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:04:56 crc kubenswrapper[5035]: I1124 20:04:56.200273 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:04:56 crc kubenswrapper[5035]: E1124 20:04:56.201092 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:05:11 crc kubenswrapper[5035]: I1124 20:05:11.201059 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:05:11 crc kubenswrapper[5035]: E1124 20:05:11.202220 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:05:24 crc kubenswrapper[5035]: I1124 20:05:24.201786 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:05:24 crc kubenswrapper[5035]: E1124 20:05:24.202889 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:05:37 crc kubenswrapper[5035]: I1124 20:05:37.200455 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:05:37 crc kubenswrapper[5035]: E1124 20:05:37.201652 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:05:51 crc kubenswrapper[5035]: I1124 20:05:51.200568 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:05:51 crc kubenswrapper[5035]: E1124 20:05:51.201600 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:06:02 crc kubenswrapper[5035]: I1124 20:06:02.212012 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:06:02 crc kubenswrapper[5035]: E1124 20:06:02.213178 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:06:17 crc kubenswrapper[5035]: I1124 20:06:17.200242 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:06:17 crc kubenswrapper[5035]: E1124 20:06:17.201406 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:06:30 crc kubenswrapper[5035]: I1124 20:06:30.200204 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:06:30 crc kubenswrapper[5035]: E1124 20:06:30.201255 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:06:45 crc kubenswrapper[5035]: I1124 20:06:45.201591 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:06:45 crc kubenswrapper[5035]: E1124 20:06:45.202600 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:07:00 crc kubenswrapper[5035]: I1124 20:07:00.200595 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:07:00 crc kubenswrapper[5035]: E1124 20:07:00.201836 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:07:11 crc kubenswrapper[5035]: I1124 20:07:11.199903 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:07:11 crc kubenswrapper[5035]: E1124 20:07:11.200638 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:07:23 crc kubenswrapper[5035]: I1124 20:07:23.167621 5035 generic.go:334] "Generic (PLEG): container finished" podID="db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" containerID="6841bb91fdfa12418a914ec6ab45aff1074be057817590543f18798be685acef" exitCode=0 Nov 24 20:07:23 crc kubenswrapper[5035]: I1124 20:07:23.167715 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" event={"ID":"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d","Type":"ContainerDied","Data":"6841bb91fdfa12418a914ec6ab45aff1074be057817590543f18798be685acef"} Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.671532 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.790579 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7lhf\" (UniqueName: \"kubernetes.io/projected/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-kube-api-access-l7lhf\") pod \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.790685 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ceph\") pod \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.790707 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-combined-ca-bundle\") pod \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.790750 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ssh-key\") pod \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.790789 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-inventory\") pod \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.790875 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-secret-0\") pod \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\" (UID: \"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d\") " Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.797939 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" (UID: "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.798758 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ceph" (OuterVolumeSpecName: "ceph") pod "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" (UID: "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.799518 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-kube-api-access-l7lhf" (OuterVolumeSpecName: "kube-api-access-l7lhf") pod "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" (UID: "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d"). InnerVolumeSpecName "kube-api-access-l7lhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.818182 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-inventory" (OuterVolumeSpecName: "inventory") pod "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" (UID: "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.820949 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" (UID: "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.828497 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" (UID: "db87bf3b-6dd8-448d-80fe-8b8db1ecca5d"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.892809 5035 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.892842 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7lhf\" (UniqueName: \"kubernetes.io/projected/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-kube-api-access-l7lhf\") on node \"crc\" DevicePath \"\"" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.892857 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.892869 5035 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.892881 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 20:07:24 crc kubenswrapper[5035]: I1124 20:07:24.892891 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db87bf3b-6dd8-448d-80fe-8b8db1ecca5d-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.190225 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" event={"ID":"db87bf3b-6dd8-448d-80fe-8b8db1ecca5d","Type":"ContainerDied","Data":"3770bb5a4177d426ebee32563d95b6723b89f0a94cdf0f65bdc5a16c334bd62d"} Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.190285 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3770bb5a4177d426ebee32563d95b6723b89f0a94cdf0f65bdc5a16c334bd62d" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.190374 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fj957" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.200107 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:07:25 crc kubenswrapper[5035]: E1124 20:07:25.200493 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.335630 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx"] Nov 24 20:07:25 crc kubenswrapper[5035]: E1124 20:07:25.336600 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.336822 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.337441 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="db87bf3b-6dd8-448d-80fe-8b8db1ecca5d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.338614 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.342075 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-r7dbz" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.342366 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.342504 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.342922 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.343740 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.343778 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.343957 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.343975 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.345247 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.349433 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx"] Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505314 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505363 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505407 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505520 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505597 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtch5\" (UniqueName: \"kubernetes.io/projected/4bbedfca-6249-484a-a6cf-da8b54a51b1e-kube-api-access-wtch5\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505621 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505741 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505769 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505809 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.505895 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.506019 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.608223 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.608464 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.608600 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.608711 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.608781 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtch5\" (UniqueName: \"kubernetes.io/projected/4bbedfca-6249-484a-a6cf-da8b54a51b1e-kube-api-access-wtch5\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.608844 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.608968 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.609024 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.609086 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.609180 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.609274 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.610629 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.612151 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.616159 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.616342 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.616743 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.617195 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.618015 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.618427 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.619024 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.623639 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.641232 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtch5\" (UniqueName: \"kubernetes.io/projected/4bbedfca-6249-484a-a6cf-da8b54a51b1e-kube-api-access-wtch5\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:25 crc kubenswrapper[5035]: I1124 20:07:25.671191 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:07:26 crc kubenswrapper[5035]: I1124 20:07:26.303439 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx"] Nov 24 20:07:26 crc kubenswrapper[5035]: I1124 20:07:26.309024 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:07:27 crc kubenswrapper[5035]: I1124 20:07:27.218812 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" event={"ID":"4bbedfca-6249-484a-a6cf-da8b54a51b1e","Type":"ContainerStarted","Data":"ece46de4fb2c3835d655997aa35ef78c56cb1981a269790f6aea9adb1be2cabe"} Nov 24 20:07:27 crc kubenswrapper[5035]: I1124 20:07:27.219930 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" event={"ID":"4bbedfca-6249-484a-a6cf-da8b54a51b1e","Type":"ContainerStarted","Data":"ab8ae57099066da72d1d7d70e2526d17b9fe60411d34a8bcbe049df249e3782d"} Nov 24 20:07:27 crc kubenswrapper[5035]: I1124 20:07:27.242896 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" podStartSLOduration=1.6544369030000001 podStartE2EDuration="2.242870214s" podCreationTimestamp="2025-11-24 20:07:25 +0000 UTC" firstStartedPulling="2025-11-24 20:07:26.308843906 +0000 UTC m=+3184.831350163" lastFinishedPulling="2025-11-24 20:07:26.897277207 +0000 UTC m=+3185.419783474" observedRunningTime="2025-11-24 20:07:27.240028087 +0000 UTC m=+3185.762534344" watchObservedRunningTime="2025-11-24 20:07:27.242870214 +0000 UTC m=+3185.765376501" Nov 24 20:07:37 crc kubenswrapper[5035]: I1124 20:07:37.200627 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:07:37 crc kubenswrapper[5035]: E1124 20:07:37.202677 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:07:49 crc kubenswrapper[5035]: I1124 20:07:49.200838 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:07:49 crc kubenswrapper[5035]: E1124 20:07:49.201866 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:08:02 crc kubenswrapper[5035]: I1124 20:08:02.208775 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:08:02 crc kubenswrapper[5035]: E1124 20:08:02.211235 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:08:13 crc kubenswrapper[5035]: I1124 20:08:13.200400 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:08:13 crc kubenswrapper[5035]: E1124 20:08:13.201211 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:08:24 crc kubenswrapper[5035]: I1124 20:08:24.201174 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:08:24 crc kubenswrapper[5035]: I1124 20:08:24.862573 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"cbcdbc5dae82463d722971b6c0e1597afa8425f378dbbe8431940fc4c08a86af"} Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.539625 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qxrkx"] Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.544279 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.565449 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qxrkx"] Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.737411 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cpkc\" (UniqueName: \"kubernetes.io/projected/d2761917-efbe-457e-9b13-b9298d23ee9e-kube-api-access-8cpkc\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.737460 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-catalog-content\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.737513 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-utilities\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.840313 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-utilities\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.840790 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cpkc\" (UniqueName: \"kubernetes.io/projected/d2761917-efbe-457e-9b13-b9298d23ee9e-kube-api-access-8cpkc\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.840820 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-catalog-content\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.840847 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-utilities\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.841074 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-catalog-content\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:53 crc kubenswrapper[5035]: I1124 20:09:53.870502 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cpkc\" (UniqueName: \"kubernetes.io/projected/d2761917-efbe-457e-9b13-b9298d23ee9e-kube-api-access-8cpkc\") pod \"redhat-operators-qxrkx\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:54 crc kubenswrapper[5035]: I1124 20:09:54.163255 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:09:54 crc kubenswrapper[5035]: I1124 20:09:54.679109 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qxrkx"] Nov 24 20:09:54 crc kubenswrapper[5035]: I1124 20:09:54.928211 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qxrkx" event={"ID":"d2761917-efbe-457e-9b13-b9298d23ee9e","Type":"ContainerStarted","Data":"657154b82f9dbaff6e2850be12a0572c1e0c2c44d6a68c4880ed71593e044c28"} Nov 24 20:09:55 crc kubenswrapper[5035]: I1124 20:09:55.938962 5035 generic.go:334] "Generic (PLEG): container finished" podID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerID="bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5" exitCode=0 Nov 24 20:09:55 crc kubenswrapper[5035]: I1124 20:09:55.939088 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qxrkx" event={"ID":"d2761917-efbe-457e-9b13-b9298d23ee9e","Type":"ContainerDied","Data":"bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5"} Nov 24 20:09:57 crc kubenswrapper[5035]: I1124 20:09:57.965697 5035 generic.go:334] "Generic (PLEG): container finished" podID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerID="36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f" exitCode=0 Nov 24 20:09:57 crc kubenswrapper[5035]: I1124 20:09:57.965827 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qxrkx" event={"ID":"d2761917-efbe-457e-9b13-b9298d23ee9e","Type":"ContainerDied","Data":"36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f"} Nov 24 20:09:59 crc kubenswrapper[5035]: I1124 20:09:59.990718 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qxrkx" event={"ID":"d2761917-efbe-457e-9b13-b9298d23ee9e","Type":"ContainerStarted","Data":"9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb"} Nov 24 20:10:00 crc kubenswrapper[5035]: I1124 20:10:00.024804 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qxrkx" podStartSLOduration=4.205475726 podStartE2EDuration="7.024780675s" podCreationTimestamp="2025-11-24 20:09:53 +0000 UTC" firstStartedPulling="2025-11-24 20:09:55.94249187 +0000 UTC m=+3334.464998177" lastFinishedPulling="2025-11-24 20:09:58.761796829 +0000 UTC m=+3337.284303126" observedRunningTime="2025-11-24 20:10:00.017719944 +0000 UTC m=+3338.540226211" watchObservedRunningTime="2025-11-24 20:10:00.024780675 +0000 UTC m=+3338.547286942" Nov 24 20:10:04 crc kubenswrapper[5035]: I1124 20:10:04.164300 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:10:04 crc kubenswrapper[5035]: I1124 20:10:04.164831 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:10:05 crc kubenswrapper[5035]: I1124 20:10:05.221638 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qxrkx" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="registry-server" probeResult="failure" output=< Nov 24 20:10:05 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:10:05 crc kubenswrapper[5035]: > Nov 24 20:10:14 crc kubenswrapper[5035]: I1124 20:10:14.234273 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:10:14 crc kubenswrapper[5035]: I1124 20:10:14.300805 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:10:14 crc kubenswrapper[5035]: I1124 20:10:14.483278 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qxrkx"] Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.160683 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qxrkx" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="registry-server" containerID="cri-o://9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb" gracePeriod=2 Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.711496 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.842871 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cpkc\" (UniqueName: \"kubernetes.io/projected/d2761917-efbe-457e-9b13-b9298d23ee9e-kube-api-access-8cpkc\") pod \"d2761917-efbe-457e-9b13-b9298d23ee9e\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.842976 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-catalog-content\") pod \"d2761917-efbe-457e-9b13-b9298d23ee9e\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.843030 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-utilities\") pod \"d2761917-efbe-457e-9b13-b9298d23ee9e\" (UID: \"d2761917-efbe-457e-9b13-b9298d23ee9e\") " Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.844148 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-utilities" (OuterVolumeSpecName: "utilities") pod "d2761917-efbe-457e-9b13-b9298d23ee9e" (UID: "d2761917-efbe-457e-9b13-b9298d23ee9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.851656 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2761917-efbe-457e-9b13-b9298d23ee9e-kube-api-access-8cpkc" (OuterVolumeSpecName: "kube-api-access-8cpkc") pod "d2761917-efbe-457e-9b13-b9298d23ee9e" (UID: "d2761917-efbe-457e-9b13-b9298d23ee9e"). InnerVolumeSpecName "kube-api-access-8cpkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.929570 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2761917-efbe-457e-9b13-b9298d23ee9e" (UID: "d2761917-efbe-457e-9b13-b9298d23ee9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.945316 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cpkc\" (UniqueName: \"kubernetes.io/projected/d2761917-efbe-457e-9b13-b9298d23ee9e-kube-api-access-8cpkc\") on node \"crc\" DevicePath \"\"" Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.945354 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:10:16 crc kubenswrapper[5035]: I1124 20:10:16.945364 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2761917-efbe-457e-9b13-b9298d23ee9e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.177554 5035 generic.go:334] "Generic (PLEG): container finished" podID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerID="9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb" exitCode=0 Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.177725 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qxrkx" event={"ID":"d2761917-efbe-457e-9b13-b9298d23ee9e","Type":"ContainerDied","Data":"9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb"} Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.177748 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qxrkx" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.178132 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qxrkx" event={"ID":"d2761917-efbe-457e-9b13-b9298d23ee9e","Type":"ContainerDied","Data":"657154b82f9dbaff6e2850be12a0572c1e0c2c44d6a68c4880ed71593e044c28"} Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.178181 5035 scope.go:117] "RemoveContainer" containerID="9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.243959 5035 scope.go:117] "RemoveContainer" containerID="36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.245869 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qxrkx"] Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.261169 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qxrkx"] Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.283934 5035 scope.go:117] "RemoveContainer" containerID="bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.312547 5035 scope.go:117] "RemoveContainer" containerID="9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb" Nov 24 20:10:17 crc kubenswrapper[5035]: E1124 20:10:17.313126 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb\": container with ID starting with 9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb not found: ID does not exist" containerID="9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.313191 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb"} err="failed to get container status \"9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb\": rpc error: code = NotFound desc = could not find container \"9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb\": container with ID starting with 9ed5282b241c2e22a10bb49f5386be9d878b18fc88ffeb1469074d2f7efabecb not found: ID does not exist" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.313222 5035 scope.go:117] "RemoveContainer" containerID="36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f" Nov 24 20:10:17 crc kubenswrapper[5035]: E1124 20:10:17.313793 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f\": container with ID starting with 36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f not found: ID does not exist" containerID="36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.313821 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f"} err="failed to get container status \"36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f\": rpc error: code = NotFound desc = could not find container \"36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f\": container with ID starting with 36de4b4c00104ff6cbd9543a94128f1b311d92c6a3a4918d1b5f3cd64536321f not found: ID does not exist" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.313847 5035 scope.go:117] "RemoveContainer" containerID="bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5" Nov 24 20:10:17 crc kubenswrapper[5035]: E1124 20:10:17.314151 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5\": container with ID starting with bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5 not found: ID does not exist" containerID="bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5" Nov 24 20:10:17 crc kubenswrapper[5035]: I1124 20:10:17.314201 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5"} err="failed to get container status \"bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5\": rpc error: code = NotFound desc = could not find container \"bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5\": container with ID starting with bf4dda8ea35d8cab411d8c59720690b015c7a2255d0fd76cfd4e5051a6a31ea5 not found: ID does not exist" Nov 24 20:10:18 crc kubenswrapper[5035]: I1124 20:10:18.218959 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" path="/var/lib/kubelet/pods/d2761917-efbe-457e-9b13-b9298d23ee9e/volumes" Nov 24 20:10:45 crc kubenswrapper[5035]: I1124 20:10:45.234353 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:10:45 crc kubenswrapper[5035]: I1124 20:10:45.235164 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:11:02 crc kubenswrapper[5035]: I1124 20:11:02.651948 5035 generic.go:334] "Generic (PLEG): container finished" podID="4bbedfca-6249-484a-a6cf-da8b54a51b1e" containerID="ece46de4fb2c3835d655997aa35ef78c56cb1981a269790f6aea9adb1be2cabe" exitCode=0 Nov 24 20:11:02 crc kubenswrapper[5035]: I1124 20:11:02.652038 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" event={"ID":"4bbedfca-6249-484a-a6cf-da8b54a51b1e","Type":"ContainerDied","Data":"ece46de4fb2c3835d655997aa35ef78c56cb1981a269790f6aea9adb1be2cabe"} Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.219941 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.375489 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtch5\" (UniqueName: \"kubernetes.io/projected/4bbedfca-6249-484a-a6cf-da8b54a51b1e-kube-api-access-wtch5\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.375603 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ssh-key\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.375649 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-1\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.375751 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-custom-ceph-combined-ca-bundle\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.375839 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-0\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.375890 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.376547 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-0\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.376663 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-inventory\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.376693 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-extra-config-0\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.376734 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph-nova-0\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.376809 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-1\") pod \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\" (UID: \"4bbedfca-6249-484a-a6cf-da8b54a51b1e\") " Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.383197 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbedfca-6249-484a-a6cf-da8b54a51b1e-kube-api-access-wtch5" (OuterVolumeSpecName: "kube-api-access-wtch5") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "kube-api-access-wtch5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.397925 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.399928 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph" (OuterVolumeSpecName: "ceph") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.404947 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.406040 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-inventory" (OuterVolumeSpecName: "inventory") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.409784 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.418704 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.420870 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.423592 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.424823 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.432125 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4bbedfca-6249-484a-a6cf-da8b54a51b1e" (UID: "4bbedfca-6249-484a-a6cf-da8b54a51b1e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480482 5035 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480551 5035 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480564 5035 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480574 5035 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480585 5035 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480595 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtch5\" (UniqueName: \"kubernetes.io/projected/4bbedfca-6249-484a-a6cf-da8b54a51b1e-kube-api-access-wtch5\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480604 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480614 5035 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480624 5035 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480635 5035 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.480645 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4bbedfca-6249-484a-a6cf-da8b54a51b1e-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.678174 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" event={"ID":"4bbedfca-6249-484a-a6cf-da8b54a51b1e","Type":"ContainerDied","Data":"ab8ae57099066da72d1d7d70e2526d17b9fe60411d34a8bcbe049df249e3782d"} Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.678240 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab8ae57099066da72d1d7d70e2526d17b9fe60411d34a8bcbe049df249e3782d" Nov 24 20:11:04 crc kubenswrapper[5035]: I1124 20:11:04.678410 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.511583 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5m8ds"] Nov 24 20:11:07 crc kubenswrapper[5035]: E1124 20:11:07.513585 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="extract-content" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.513713 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="extract-content" Nov 24 20:11:07 crc kubenswrapper[5035]: E1124 20:11:07.513817 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="registry-server" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.513894 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="registry-server" Nov 24 20:11:07 crc kubenswrapper[5035]: E1124 20:11:07.513988 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbedfca-6249-484a-a6cf-da8b54a51b1e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.514071 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbedfca-6249-484a-a6cf-da8b54a51b1e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 24 20:11:07 crc kubenswrapper[5035]: E1124 20:11:07.514155 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="extract-utilities" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.514234 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="extract-utilities" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.514561 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbedfca-6249-484a-a6cf-da8b54a51b1e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.514663 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2761917-efbe-457e-9b13-b9298d23ee9e" containerName="registry-server" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.516798 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.535185 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m8ds"] Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.646145 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-utilities\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.646223 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-catalog-content\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.646349 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l6t9\" (UniqueName: \"kubernetes.io/projected/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-kube-api-access-6l6t9\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.747436 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-utilities\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.747495 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-catalog-content\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.747595 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l6t9\" (UniqueName: \"kubernetes.io/projected/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-kube-api-access-6l6t9\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.747982 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-utilities\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.748200 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-catalog-content\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.767091 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l6t9\" (UniqueName: \"kubernetes.io/projected/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-kube-api-access-6l6t9\") pod \"redhat-marketplace-5m8ds\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:07 crc kubenswrapper[5035]: I1124 20:11:07.910232 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:08 crc kubenswrapper[5035]: I1124 20:11:08.355702 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m8ds"] Nov 24 20:11:08 crc kubenswrapper[5035]: I1124 20:11:08.731264 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m8ds" event={"ID":"d2ee84bd-a6e4-4328-8500-7eda2ca113b8","Type":"ContainerStarted","Data":"4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d"} Nov 24 20:11:08 crc kubenswrapper[5035]: I1124 20:11:08.731860 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m8ds" event={"ID":"d2ee84bd-a6e4-4328-8500-7eda2ca113b8","Type":"ContainerStarted","Data":"506452be5cda255cbb8b830c4007fd27702544e4cab7f63da1f0bffc72b17641"} Nov 24 20:11:09 crc kubenswrapper[5035]: I1124 20:11:09.743336 5035 generic.go:334] "Generic (PLEG): container finished" podID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerID="4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d" exitCode=0 Nov 24 20:11:09 crc kubenswrapper[5035]: I1124 20:11:09.743440 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m8ds" event={"ID":"d2ee84bd-a6e4-4328-8500-7eda2ca113b8","Type":"ContainerDied","Data":"4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d"} Nov 24 20:11:11 crc kubenswrapper[5035]: I1124 20:11:11.767803 5035 generic.go:334] "Generic (PLEG): container finished" podID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerID="2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374" exitCode=0 Nov 24 20:11:11 crc kubenswrapper[5035]: I1124 20:11:11.767864 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m8ds" event={"ID":"d2ee84bd-a6e4-4328-8500-7eda2ca113b8","Type":"ContainerDied","Data":"2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374"} Nov 24 20:11:12 crc kubenswrapper[5035]: I1124 20:11:12.777405 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m8ds" event={"ID":"d2ee84bd-a6e4-4328-8500-7eda2ca113b8","Type":"ContainerStarted","Data":"f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8"} Nov 24 20:11:12 crc kubenswrapper[5035]: I1124 20:11:12.798003 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5m8ds" podStartSLOduration=3.217279819 podStartE2EDuration="5.797982458s" podCreationTimestamp="2025-11-24 20:11:07 +0000 UTC" firstStartedPulling="2025-11-24 20:11:09.757231664 +0000 UTC m=+3408.279737931" lastFinishedPulling="2025-11-24 20:11:12.337934283 +0000 UTC m=+3410.860440570" observedRunningTime="2025-11-24 20:11:12.792848751 +0000 UTC m=+3411.315355008" watchObservedRunningTime="2025-11-24 20:11:12.797982458 +0000 UTC m=+3411.320488725" Nov 24 20:11:15 crc kubenswrapper[5035]: I1124 20:11:15.233839 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:11:15 crc kubenswrapper[5035]: I1124 20:11:15.234235 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:11:17 crc kubenswrapper[5035]: I1124 20:11:17.911807 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:17 crc kubenswrapper[5035]: I1124 20:11:17.912189 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:17 crc kubenswrapper[5035]: I1124 20:11:17.972284 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:18 crc kubenswrapper[5035]: I1124 20:11:18.932422 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.000889 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m8ds"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.212578 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.213978 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.216360 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.216535 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.227158 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.234880 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.234973 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.256713 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.274005 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302504 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302556 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302589 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302619 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wznng\" (UniqueName: \"kubernetes.io/projected/4d0bc83b-3e60-45a1-a251-06a00909de0a-kube-api-access-wznng\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302641 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302691 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302708 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302725 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-sys\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302744 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302758 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302807 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302823 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302837 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-config-data\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302852 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302871 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-run\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302887 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-lib-modules\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302915 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302930 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302950 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fc314ad6-dfe5-4703-9f08-f08f904b605c-ceph\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302964 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302977 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.302992 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-scripts\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303011 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-dev\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303026 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303054 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303071 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303084 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4d0bc83b-3e60-45a1-a251-06a00909de0a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303115 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303144 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-run\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303190 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znf9b\" (UniqueName: \"kubernetes.io/projected/fc314ad6-dfe5-4703-9f08-f08f904b605c-kube-api-access-znf9b\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303222 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.303238 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405028 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405075 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405102 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fc314ad6-dfe5-4703-9f08-f08f904b605c-ceph\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405138 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405155 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405174 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-scripts\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405192 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-dev\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405209 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405234 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405250 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4d0bc83b-3e60-45a1-a251-06a00909de0a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405263 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405309 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405319 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405417 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-run\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405466 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znf9b\" (UniqueName: \"kubernetes.io/projected/fc314ad6-dfe5-4703-9f08-f08f904b605c-kube-api-access-znf9b\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405514 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405524 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405553 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405592 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-run\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405597 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405616 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405536 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405665 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405683 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405708 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405737 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wznng\" (UniqueName: \"kubernetes.io/projected/4d0bc83b-3e60-45a1-a251-06a00909de0a-kube-api-access-wznng\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405759 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405788 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405792 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405806 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405823 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-sys\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405843 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405859 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405893 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405908 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405933 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-config-data\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405950 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405971 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-run\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.405988 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-lib-modules\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.406049 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-lib-modules\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.406112 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.406992 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407205 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407251 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407274 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407336 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-sys\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407360 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-dev\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407379 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-run\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407301 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407417 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc314ad6-dfe5-4703-9f08-f08f904b605c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.407446 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4d0bc83b-3e60-45a1-a251-06a00909de0a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.410467 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-config-data\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.410981 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.411705 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fc314ad6-dfe5-4703-9f08-f08f904b605c-ceph\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.411811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.412235 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.412342 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4d0bc83b-3e60-45a1-a251-06a00909de0a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.415090 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.418044 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.423370 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc314ad6-dfe5-4703-9f08-f08f904b605c-scripts\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.424448 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d0bc83b-3e60-45a1-a251-06a00909de0a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.427618 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znf9b\" (UniqueName: \"kubernetes.io/projected/fc314ad6-dfe5-4703-9f08-f08f904b605c-kube-api-access-znf9b\") pod \"cinder-backup-0\" (UID: \"fc314ad6-dfe5-4703-9f08-f08f904b605c\") " pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.441210 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wznng\" (UniqueName: \"kubernetes.io/projected/4d0bc83b-3e60-45a1-a251-06a00909de0a-kube-api-access-wznng\") pod \"cinder-volume-volume1-0\" (UID: \"4d0bc83b-3e60-45a1-a251-06a00909de0a\") " pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.568148 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.576700 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.705543 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-vk6mj"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.711026 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.719843 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-vk6mj"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.793334 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-65b2-account-create-xpmrl"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.794572 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.799533 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.808077 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-65b2-account-create-xpmrl"] Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.813310 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g97lh\" (UniqueName: \"kubernetes.io/projected/7728bae7-82b0-4cf4-824f-4911ff936ed1-kube-api-access-g97lh\") pod \"manila-db-create-vk6mj\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.813361 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qdq7\" (UniqueName: \"kubernetes.io/projected/7097c70e-55c5-4d34-88a6-fce197f78ba8-kube-api-access-5qdq7\") pod \"manila-65b2-account-create-xpmrl\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.813380 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7097c70e-55c5-4d34-88a6-fce197f78ba8-operator-scripts\") pod \"manila-65b2-account-create-xpmrl\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.813442 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7728bae7-82b0-4cf4-824f-4911ff936ed1-operator-scripts\") pod \"manila-db-create-vk6mj\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.918765 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qdq7\" (UniqueName: \"kubernetes.io/projected/7097c70e-55c5-4d34-88a6-fce197f78ba8-kube-api-access-5qdq7\") pod \"manila-65b2-account-create-xpmrl\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.918826 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7097c70e-55c5-4d34-88a6-fce197f78ba8-operator-scripts\") pod \"manila-65b2-account-create-xpmrl\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.918919 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7728bae7-82b0-4cf4-824f-4911ff936ed1-operator-scripts\") pod \"manila-db-create-vk6mj\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.919066 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g97lh\" (UniqueName: \"kubernetes.io/projected/7728bae7-82b0-4cf4-824f-4911ff936ed1-kube-api-access-g97lh\") pod \"manila-db-create-vk6mj\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.920369 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7097c70e-55c5-4d34-88a6-fce197f78ba8-operator-scripts\") pod \"manila-65b2-account-create-xpmrl\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.920867 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7728bae7-82b0-4cf4-824f-4911ff936ed1-operator-scripts\") pod \"manila-db-create-vk6mj\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.936181 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qdq7\" (UniqueName: \"kubernetes.io/projected/7097c70e-55c5-4d34-88a6-fce197f78ba8-kube-api-access-5qdq7\") pod \"manila-65b2-account-create-xpmrl\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:19 crc kubenswrapper[5035]: I1124 20:11:19.936414 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g97lh\" (UniqueName: \"kubernetes.io/projected/7728bae7-82b0-4cf4-824f-4911ff936ed1-kube-api-access-g97lh\") pod \"manila-db-create-vk6mj\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.044535 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.116933 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.119032 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.122765 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.122965 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.123340 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.123368 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-r79bb" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.123987 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.133050 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.191763 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.194180 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.196640 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.200459 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.224158 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225414 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7851793-a562-491b-8c0e-47b9155a75bc-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225460 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-scripts\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225477 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225499 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/815bccaf-12a3-47b8-8e80-3c1c30724697-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225519 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225549 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7851793-a562-491b-8c0e-47b9155a75bc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225567 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225591 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/815bccaf-12a3-47b8-8e80-3c1c30724697-logs\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225682 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225706 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225732 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225769 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/815bccaf-12a3-47b8-8e80-3c1c30724697-ceph\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225785 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mjpm\" (UniqueName: \"kubernetes.io/projected/e7851793-a562-491b-8c0e-47b9155a75bc-kube-api-access-9mjpm\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225801 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-config-data\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225860 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2c7m\" (UniqueName: \"kubernetes.io/projected/815bccaf-12a3-47b8-8e80-3c1c30724697-kube-api-access-g2c7m\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225877 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.225893 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7851793-a562-491b-8c0e-47b9155a75bc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.249928 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331167 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2c7m\" (UniqueName: \"kubernetes.io/projected/815bccaf-12a3-47b8-8e80-3c1c30724697-kube-api-access-g2c7m\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331226 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331248 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7851793-a562-491b-8c0e-47b9155a75bc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331280 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7851793-a562-491b-8c0e-47b9155a75bc-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331334 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-scripts\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331353 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331384 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/815bccaf-12a3-47b8-8e80-3c1c30724697-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331416 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7851793-a562-491b-8c0e-47b9155a75bc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331473 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331498 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/815bccaf-12a3-47b8-8e80-3c1c30724697-logs\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331553 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331572 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331595 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331619 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331664 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/815bccaf-12a3-47b8-8e80-3c1c30724697-ceph\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331683 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mjpm\" (UniqueName: \"kubernetes.io/projected/e7851793-a562-491b-8c0e-47b9155a75bc-kube-api-access-9mjpm\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.331705 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-config-data\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.332380 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/815bccaf-12a3-47b8-8e80-3c1c30724697-logs\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.332498 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.332580 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.332693 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7851793-a562-491b-8c0e-47b9155a75bc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.333141 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/815bccaf-12a3-47b8-8e80-3c1c30724697-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.333351 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7851793-a562-491b-8c0e-47b9155a75bc-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.345027 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.345723 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7851793-a562-491b-8c0e-47b9155a75bc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.348045 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.349057 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-config-data\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.360474 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.363902 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.373314 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/815bccaf-12a3-47b8-8e80-3c1c30724697-ceph\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.375137 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.375361 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2c7m\" (UniqueName: \"kubernetes.io/projected/815bccaf-12a3-47b8-8e80-3c1c30724697-kube-api-access-g2c7m\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.375809 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815bccaf-12a3-47b8-8e80-3c1c30724697-scripts\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.376779 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7851793-a562-491b-8c0e-47b9155a75bc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.378542 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mjpm\" (UniqueName: \"kubernetes.io/projected/e7851793-a562-491b-8c0e-47b9155a75bc-kube-api-access-9mjpm\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.381714 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"815bccaf-12a3-47b8-8e80-3c1c30724697\") " pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.399827 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7851793-a562-491b-8c0e-47b9155a75bc\") " pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.455428 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.526228 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.598934 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-65b2-account-create-xpmrl"] Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.729342 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-vk6mj"] Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.875548 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-65b2-account-create-xpmrl" event={"ID":"7097c70e-55c5-4d34-88a6-fce197f78ba8","Type":"ContainerStarted","Data":"ccc9b0c7d620c2f199b54c99bf5bebe87b0542142fb7e9cbf1dfb768804c4486"} Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.876721 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-vk6mj" event={"ID":"7728bae7-82b0-4cf4-824f-4911ff936ed1","Type":"ContainerStarted","Data":"b33031d2ca7d8e0fc17959ea7287ca9e788de55b19aac5da0b6a420059f85ca1"} Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.878709 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"fc314ad6-dfe5-4703-9f08-f08f904b605c","Type":"ContainerStarted","Data":"ec608c76d82265bf4a518ff00aeb9eb95e08fad73f9119da5e9279c7642ab533"} Nov 24 20:11:20 crc kubenswrapper[5035]: I1124 20:11:20.878815 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5m8ds" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="registry-server" containerID="cri-o://f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8" gracePeriod=2 Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.248339 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 20:11:21 crc kubenswrapper[5035]: W1124 20:11:21.288721 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d0bc83b_3e60_45a1_a251_06a00909de0a.slice/crio-1c0ef682647a64929c8a0d855fe4526ba9c939d12e002e8948eef6d704127a6e WatchSource:0}: Error finding container 1c0ef682647a64929c8a0d855fe4526ba9c939d12e002e8948eef6d704127a6e: Status 404 returned error can't find the container with id 1c0ef682647a64929c8a0d855fe4526ba9c939d12e002e8948eef6d704127a6e Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.293262 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.340405 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 20:11:21 crc kubenswrapper[5035]: W1124 20:11:21.382480 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7851793_a562_491b_8c0e_47b9155a75bc.slice/crio-f9ce4ceb731f456ecc7f2d8a72694e029f36d58b5c2781674675993efb0c3602 WatchSource:0}: Error finding container f9ce4ceb731f456ecc7f2d8a72694e029f36d58b5c2781674675993efb0c3602: Status 404 returned error can't find the container with id f9ce4ceb731f456ecc7f2d8a72694e029f36d58b5c2781674675993efb0c3602 Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.391439 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l6t9\" (UniqueName: \"kubernetes.io/projected/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-kube-api-access-6l6t9\") pod \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.391598 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-utilities\") pod \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.391639 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-catalog-content\") pod \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\" (UID: \"d2ee84bd-a6e4-4328-8500-7eda2ca113b8\") " Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.392644 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-utilities" (OuterVolumeSpecName: "utilities") pod "d2ee84bd-a6e4-4328-8500-7eda2ca113b8" (UID: "d2ee84bd-a6e4-4328-8500-7eda2ca113b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.398339 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-kube-api-access-6l6t9" (OuterVolumeSpecName: "kube-api-access-6l6t9") pod "d2ee84bd-a6e4-4328-8500-7eda2ca113b8" (UID: "d2ee84bd-a6e4-4328-8500-7eda2ca113b8"). InnerVolumeSpecName "kube-api-access-6l6t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.412225 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2ee84bd-a6e4-4328-8500-7eda2ca113b8" (UID: "d2ee84bd-a6e4-4328-8500-7eda2ca113b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.493638 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l6t9\" (UniqueName: \"kubernetes.io/projected/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-kube-api-access-6l6t9\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.493969 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.493981 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2ee84bd-a6e4-4328-8500-7eda2ca113b8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.889432 5035 generic.go:334] "Generic (PLEG): container finished" podID="7097c70e-55c5-4d34-88a6-fce197f78ba8" containerID="7aba46386f064f79f23c934a770724b2b6674d54aeba739c9a8f18451a3ba143" exitCode=0 Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.889756 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-65b2-account-create-xpmrl" event={"ID":"7097c70e-55c5-4d34-88a6-fce197f78ba8","Type":"ContainerDied","Data":"7aba46386f064f79f23c934a770724b2b6674d54aeba739c9a8f18451a3ba143"} Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.892206 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"4d0bc83b-3e60-45a1-a251-06a00909de0a","Type":"ContainerStarted","Data":"1c0ef682647a64929c8a0d855fe4526ba9c939d12e002e8948eef6d704127a6e"} Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.894569 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7851793-a562-491b-8c0e-47b9155a75bc","Type":"ContainerStarted","Data":"f9ce4ceb731f456ecc7f2d8a72694e029f36d58b5c2781674675993efb0c3602"} Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.897240 5035 generic.go:334] "Generic (PLEG): container finished" podID="7728bae7-82b0-4cf4-824f-4911ff936ed1" containerID="ce26f55c82c86ec494bf4df0fec0dd84d40355744726fcaa595220158601bbdd" exitCode=0 Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.897326 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-vk6mj" event={"ID":"7728bae7-82b0-4cf4-824f-4911ff936ed1","Type":"ContainerDied","Data":"ce26f55c82c86ec494bf4df0fec0dd84d40355744726fcaa595220158601bbdd"} Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.906796 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"fc314ad6-dfe5-4703-9f08-f08f904b605c","Type":"ContainerStarted","Data":"42c08223a01b3289eee9b2a0e92c6c3d0974844d059cf649db49925f09905c99"} Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.933221 5035 generic.go:334] "Generic (PLEG): container finished" podID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerID="f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8" exitCode=0 Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.933266 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m8ds" event={"ID":"d2ee84bd-a6e4-4328-8500-7eda2ca113b8","Type":"ContainerDied","Data":"f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8"} Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.933302 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m8ds" event={"ID":"d2ee84bd-a6e4-4328-8500-7eda2ca113b8","Type":"ContainerDied","Data":"506452be5cda255cbb8b830c4007fd27702544e4cab7f63da1f0bffc72b17641"} Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.933312 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m8ds" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.933318 5035 scope.go:117] "RemoveContainer" containerID="f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.978649 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m8ds"] Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.980138 5035 scope.go:117] "RemoveContainer" containerID="2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374" Nov 24 20:11:21 crc kubenswrapper[5035]: I1124 20:11:21.993629 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m8ds"] Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.002544 5035 scope.go:117] "RemoveContainer" containerID="4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.048035 5035 scope.go:117] "RemoveContainer" containerID="f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8" Nov 24 20:11:22 crc kubenswrapper[5035]: E1124 20:11:22.048416 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8\": container with ID starting with f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8 not found: ID does not exist" containerID="f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.048450 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8"} err="failed to get container status \"f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8\": rpc error: code = NotFound desc = could not find container \"f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8\": container with ID starting with f437b5d61e05fb81a4d86fd27447537d26cc4ee69de969a6947ee3d28d35c4a8 not found: ID does not exist" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.048471 5035 scope.go:117] "RemoveContainer" containerID="2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374" Nov 24 20:11:22 crc kubenswrapper[5035]: E1124 20:11:22.048742 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374\": container with ID starting with 2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374 not found: ID does not exist" containerID="2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.048794 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374"} err="failed to get container status \"2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374\": rpc error: code = NotFound desc = could not find container \"2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374\": container with ID starting with 2b022ff934faf00584bc9d898d1ed174e5dd0188caa16f209c9883eed371c374 not found: ID does not exist" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.048824 5035 scope.go:117] "RemoveContainer" containerID="4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d" Nov 24 20:11:22 crc kubenswrapper[5035]: E1124 20:11:22.049060 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d\": container with ID starting with 4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d not found: ID does not exist" containerID="4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.049080 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d"} err="failed to get container status \"4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d\": rpc error: code = NotFound desc = could not find container \"4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d\": container with ID starting with 4743424dc94c898f3105a5ed041483e166ef45bf846ca5f79d7d4ecb6e9b1e2d not found: ID does not exist" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.219368 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" path="/var/lib/kubelet/pods/d2ee84bd-a6e4-4328-8500-7eda2ca113b8/volumes" Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.290204 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.944840 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7851793-a562-491b-8c0e-47b9155a75bc","Type":"ContainerStarted","Data":"72e1a539af588a664b5f876a44f77bc75f28d478a865a8c4500c2be4a9284db7"} Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.945135 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7851793-a562-491b-8c0e-47b9155a75bc","Type":"ContainerStarted","Data":"826fb657f8932cbdc5534c40a5584f9ec3550954f48a852c71dea0966f0b7dc4"} Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.948426 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"fc314ad6-dfe5-4703-9f08-f08f904b605c","Type":"ContainerStarted","Data":"1cd4c65f6e175504cf121310ae3a8908750a9f38486f86e744814621786159ca"} Nov 24 20:11:22 crc kubenswrapper[5035]: I1124 20:11:22.949937 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"815bccaf-12a3-47b8-8e80-3c1c30724697","Type":"ContainerStarted","Data":"9997b2c1ad7e90dac49ea7ef68da63cecac5d1d4d43ab1c083d6990639b57e1a"} Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.011105 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.834060167 podStartE2EDuration="4.011081712s" podCreationTimestamp="2025-11-24 20:11:19 +0000 UTC" firstStartedPulling="2025-11-24 20:11:20.252842832 +0000 UTC m=+3418.775349089" lastFinishedPulling="2025-11-24 20:11:21.429864377 +0000 UTC m=+3419.952370634" observedRunningTime="2025-11-24 20:11:23.001402151 +0000 UTC m=+3421.523908448" watchObservedRunningTime="2025-11-24 20:11:23.011081712 +0000 UTC m=+3421.533587969" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.013822 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.013803916 podStartE2EDuration="4.013803916s" podCreationTimestamp="2025-11-24 20:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:11:22.978391829 +0000 UTC m=+3421.500898106" watchObservedRunningTime="2025-11-24 20:11:23.013803916 +0000 UTC m=+3421.536310173" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.291372 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.350464 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7097c70e-55c5-4d34-88a6-fce197f78ba8-operator-scripts\") pod \"7097c70e-55c5-4d34-88a6-fce197f78ba8\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.350700 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qdq7\" (UniqueName: \"kubernetes.io/projected/7097c70e-55c5-4d34-88a6-fce197f78ba8-kube-api-access-5qdq7\") pod \"7097c70e-55c5-4d34-88a6-fce197f78ba8\" (UID: \"7097c70e-55c5-4d34-88a6-fce197f78ba8\") " Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.351829 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7097c70e-55c5-4d34-88a6-fce197f78ba8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7097c70e-55c5-4d34-88a6-fce197f78ba8" (UID: "7097c70e-55c5-4d34-88a6-fce197f78ba8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.357807 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.358310 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7097c70e-55c5-4d34-88a6-fce197f78ba8-kube-api-access-5qdq7" (OuterVolumeSpecName: "kube-api-access-5qdq7") pod "7097c70e-55c5-4d34-88a6-fce197f78ba8" (UID: "7097c70e-55c5-4d34-88a6-fce197f78ba8"). InnerVolumeSpecName "kube-api-access-5qdq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.456894 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7728bae7-82b0-4cf4-824f-4911ff936ed1-operator-scripts\") pod \"7728bae7-82b0-4cf4-824f-4911ff936ed1\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.457407 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g97lh\" (UniqueName: \"kubernetes.io/projected/7728bae7-82b0-4cf4-824f-4911ff936ed1-kube-api-access-g97lh\") pod \"7728bae7-82b0-4cf4-824f-4911ff936ed1\" (UID: \"7728bae7-82b0-4cf4-824f-4911ff936ed1\") " Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.457802 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qdq7\" (UniqueName: \"kubernetes.io/projected/7097c70e-55c5-4d34-88a6-fce197f78ba8-kube-api-access-5qdq7\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.457813 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7097c70e-55c5-4d34-88a6-fce197f78ba8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.457883 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7728bae7-82b0-4cf4-824f-4911ff936ed1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7728bae7-82b0-4cf4-824f-4911ff936ed1" (UID: "7728bae7-82b0-4cf4-824f-4911ff936ed1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.466121 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7728bae7-82b0-4cf4-824f-4911ff936ed1-kube-api-access-g97lh" (OuterVolumeSpecName: "kube-api-access-g97lh") pod "7728bae7-82b0-4cf4-824f-4911ff936ed1" (UID: "7728bae7-82b0-4cf4-824f-4911ff936ed1"). InnerVolumeSpecName "kube-api-access-g97lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.559609 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g97lh\" (UniqueName: \"kubernetes.io/projected/7728bae7-82b0-4cf4-824f-4911ff936ed1-kube-api-access-g97lh\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.559654 5035 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7728bae7-82b0-4cf4-824f-4911ff936ed1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.962821 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-65b2-account-create-xpmrl" event={"ID":"7097c70e-55c5-4d34-88a6-fce197f78ba8","Type":"ContainerDied","Data":"ccc9b0c7d620c2f199b54c99bf5bebe87b0542142fb7e9cbf1dfb768804c4486"} Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.963427 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccc9b0c7d620c2f199b54c99bf5bebe87b0542142fb7e9cbf1dfb768804c4486" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.963260 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-65b2-account-create-xpmrl" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.970318 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"4d0bc83b-3e60-45a1-a251-06a00909de0a","Type":"ContainerStarted","Data":"ed8f737352f9100c97ea6d1a9f4ab463e8cf13034746cdb7e02ce81dd3b04d73"} Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.970391 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"4d0bc83b-3e60-45a1-a251-06a00909de0a","Type":"ContainerStarted","Data":"ba6a8a86653adbc4cd591d2eeabde23f9c2b867ce252478703c08185eb7ffabd"} Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.972050 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-vk6mj" event={"ID":"7728bae7-82b0-4cf4-824f-4911ff936ed1","Type":"ContainerDied","Data":"b33031d2ca7d8e0fc17959ea7287ca9e788de55b19aac5da0b6a420059f85ca1"} Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.972090 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-vk6mj" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.972102 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b33031d2ca7d8e0fc17959ea7287ca9e788de55b19aac5da0b6a420059f85ca1" Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.973955 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"815bccaf-12a3-47b8-8e80-3c1c30724697","Type":"ContainerStarted","Data":"32f578a81b0e71c0a8dc63908a071d548356fc738aca3dbd041dc119936a3e4c"} Nov 24 20:11:23 crc kubenswrapper[5035]: I1124 20:11:23.973996 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"815bccaf-12a3-47b8-8e80-3c1c30724697","Type":"ContainerStarted","Data":"c0a4da68f69e4b1fc42b9b87b9b9acd4d9a51c596e72f0a6a942cab707b220b6"} Nov 24 20:11:24 crc kubenswrapper[5035]: I1124 20:11:24.000965 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.52374461 podStartE2EDuration="5.000942857s" podCreationTimestamp="2025-11-24 20:11:19 +0000 UTC" firstStartedPulling="2025-11-24 20:11:21.29579531 +0000 UTC m=+3419.818301577" lastFinishedPulling="2025-11-24 20:11:22.772993567 +0000 UTC m=+3421.295499824" observedRunningTime="2025-11-24 20:11:23.989912899 +0000 UTC m=+3422.512419176" watchObservedRunningTime="2025-11-24 20:11:24.000942857 +0000 UTC m=+3422.523449114" Nov 24 20:11:24 crc kubenswrapper[5035]: I1124 20:11:24.023973 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.023956808 podStartE2EDuration="5.023956808s" podCreationTimestamp="2025-11-24 20:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:11:24.019314203 +0000 UTC m=+3422.541820460" watchObservedRunningTime="2025-11-24 20:11:24.023956808 +0000 UTC m=+3422.546463065" Nov 24 20:11:24 crc kubenswrapper[5035]: I1124 20:11:24.568821 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 24 20:11:24 crc kubenswrapper[5035]: I1124 20:11:24.577056 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.141257 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-85wlr"] Nov 24 20:11:25 crc kubenswrapper[5035]: E1124 20:11:25.142035 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="registry-server" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142060 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="registry-server" Nov 24 20:11:25 crc kubenswrapper[5035]: E1124 20:11:25.142078 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="extract-utilities" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142088 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="extract-utilities" Nov 24 20:11:25 crc kubenswrapper[5035]: E1124 20:11:25.142106 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7728bae7-82b0-4cf4-824f-4911ff936ed1" containerName="mariadb-database-create" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142115 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7728bae7-82b0-4cf4-824f-4911ff936ed1" containerName="mariadb-database-create" Nov 24 20:11:25 crc kubenswrapper[5035]: E1124 20:11:25.142143 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7097c70e-55c5-4d34-88a6-fce197f78ba8" containerName="mariadb-account-create" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142152 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7097c70e-55c5-4d34-88a6-fce197f78ba8" containerName="mariadb-account-create" Nov 24 20:11:25 crc kubenswrapper[5035]: E1124 20:11:25.142174 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="extract-content" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142183 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="extract-content" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142497 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7097c70e-55c5-4d34-88a6-fce197f78ba8" containerName="mariadb-account-create" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142525 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2ee84bd-a6e4-4328-8500-7eda2ca113b8" containerName="registry-server" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.142547 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7728bae7-82b0-4cf4-824f-4911ff936ed1" containerName="mariadb-database-create" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.143282 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.146852 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.147002 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-rhsq4" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.148303 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-85wlr"] Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.298871 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmnd6\" (UniqueName: \"kubernetes.io/projected/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-kube-api-access-lmnd6\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.299435 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-config-data\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.299557 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-job-config-data\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.299625 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-combined-ca-bundle\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.401246 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmnd6\" (UniqueName: \"kubernetes.io/projected/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-kube-api-access-lmnd6\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.401555 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-config-data\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.401652 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-job-config-data\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.401737 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-combined-ca-bundle\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.413145 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-job-config-data\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.413233 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-config-data\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.414087 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-combined-ca-bundle\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.419496 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmnd6\" (UniqueName: \"kubernetes.io/projected/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-kube-api-access-lmnd6\") pod \"manila-db-sync-85wlr\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.464698 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.804998 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-85wlr"] Nov 24 20:11:25 crc kubenswrapper[5035]: I1124 20:11:25.990954 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-85wlr" event={"ID":"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25","Type":"ContainerStarted","Data":"fccfc8ffc3bff2a86c317a22a8e141580a230018c3141ff020336d3cd9042b6f"} Nov 24 20:11:29 crc kubenswrapper[5035]: I1124 20:11:29.788712 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 24 20:11:29 crc kubenswrapper[5035]: I1124 20:11:29.795910 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.457013 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.457084 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.503317 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.503799 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.532806 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.532895 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.752849 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:30 crc kubenswrapper[5035]: I1124 20:11:30.753171 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:31 crc kubenswrapper[5035]: I1124 20:11:31.045520 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 20:11:31 crc kubenswrapper[5035]: I1124 20:11:31.045560 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:31 crc kubenswrapper[5035]: I1124 20:11:31.045570 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 20:11:31 crc kubenswrapper[5035]: I1124 20:11:31.045578 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:32 crc kubenswrapper[5035]: I1124 20:11:32.066213 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-85wlr" event={"ID":"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25","Type":"ContainerStarted","Data":"c7642684733f2859e2722ec4b01e9692f1069043d5c03825dfc42e8789eb6538"} Nov 24 20:11:32 crc kubenswrapper[5035]: I1124 20:11:32.097809 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-85wlr" podStartSLOduration=2.277732632 podStartE2EDuration="7.097784927s" podCreationTimestamp="2025-11-24 20:11:25 +0000 UTC" firstStartedPulling="2025-11-24 20:11:25.809459595 +0000 UTC m=+3424.331965852" lastFinishedPulling="2025-11-24 20:11:30.62951188 +0000 UTC m=+3429.152018147" observedRunningTime="2025-11-24 20:11:32.084833386 +0000 UTC m=+3430.607339683" watchObservedRunningTime="2025-11-24 20:11:32.097784927 +0000 UTC m=+3430.620291184" Nov 24 20:11:33 crc kubenswrapper[5035]: I1124 20:11:33.074045 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 20:11:33 crc kubenswrapper[5035]: I1124 20:11:33.074389 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 20:11:33 crc kubenswrapper[5035]: I1124 20:11:33.927151 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 20:11:33 crc kubenswrapper[5035]: I1124 20:11:33.927747 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 20:11:33 crc kubenswrapper[5035]: I1124 20:11:33.938594 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:33 crc kubenswrapper[5035]: I1124 20:11:33.938701 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 20:11:33 crc kubenswrapper[5035]: I1124 20:11:33.939510 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 20:11:42 crc kubenswrapper[5035]: I1124 20:11:42.170269 5035 generic.go:334] "Generic (PLEG): container finished" podID="8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" containerID="c7642684733f2859e2722ec4b01e9692f1069043d5c03825dfc42e8789eb6538" exitCode=0 Nov 24 20:11:42 crc kubenswrapper[5035]: I1124 20:11:42.170322 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-85wlr" event={"ID":"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25","Type":"ContainerDied","Data":"c7642684733f2859e2722ec4b01e9692f1069043d5c03825dfc42e8789eb6538"} Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.669715 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.778312 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-config-data\") pod \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.778373 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-combined-ca-bundle\") pod \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.778507 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmnd6\" (UniqueName: \"kubernetes.io/projected/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-kube-api-access-lmnd6\") pod \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.778620 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-job-config-data\") pod \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\" (UID: \"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25\") " Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.785205 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" (UID: "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.785352 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-kube-api-access-lmnd6" (OuterVolumeSpecName: "kube-api-access-lmnd6") pod "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" (UID: "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25"). InnerVolumeSpecName "kube-api-access-lmnd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.787709 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-config-data" (OuterVolumeSpecName: "config-data") pod "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" (UID: "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.812111 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" (UID: "8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.880590 5035 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.880637 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.880649 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:43 crc kubenswrapper[5035]: I1124 20:11:43.880663 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmnd6\" (UniqueName: \"kubernetes.io/projected/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25-kube-api-access-lmnd6\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.193505 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-85wlr" event={"ID":"8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25","Type":"ContainerDied","Data":"fccfc8ffc3bff2a86c317a22a8e141580a230018c3141ff020336d3cd9042b6f"} Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.193554 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fccfc8ffc3bff2a86c317a22a8e141580a230018c3141ff020336d3cd9042b6f" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.193590 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-85wlr" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.592593 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:11:44 crc kubenswrapper[5035]: E1124 20:11:44.593227 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" containerName="manila-db-sync" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.593249 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" containerName="manila-db-sync" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.593525 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" containerName="manila-db-sync" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.604864 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.607810 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.607866 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.608017 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.613131 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-rhsq4" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.615332 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.708472 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.708531 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-scripts\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.708550 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-254q4\" (UniqueName: \"kubernetes.io/projected/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-kube-api-access-254q4\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.708569 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.708587 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.708614 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.723171 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.724756 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.730703 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.734426 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78f48d6b7c-zbdgc"] Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.735950 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.751091 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.768299 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78f48d6b7c-zbdgc"] Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.810825 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.810885 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.811034 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.811072 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-scripts\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.811092 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-254q4\" (UniqueName: \"kubernetes.io/projected/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-kube-api-access-254q4\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.811112 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.812278 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.816563 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.817398 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.820997 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.829170 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.833639 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.834164 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-254q4\" (UniqueName: \"kubernetes.io/projected/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-kube-api-access-254q4\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.837816 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-scripts\") pod \"manila-scheduler-0\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " pod="openstack/manila-scheduler-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.837972 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.843149 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.912867 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-scripts\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.912919 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.912949 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-ceph\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.912969 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-ovsdbserver-sb\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913000 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913100 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-openstack-edpm-ipam\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913121 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913150 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8jw9\" (UniqueName: \"kubernetes.io/projected/8d6b76db-f778-4b7e-9dab-e0fa5921458e-kube-api-access-h8jw9\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913178 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913359 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-config\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913408 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-dns-svc\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913457 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-ovsdbserver-nb\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913479 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzqd9\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-kube-api-access-bzqd9\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.913563 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:44 crc kubenswrapper[5035]: I1124 20:11:44.932689 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015492 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-config\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015538 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-dns-svc\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015566 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-ovsdbserver-nb\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015586 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzqd9\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-kube-api-access-bzqd9\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015625 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015646 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-scripts\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015678 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015713 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b7293dea-c387-4da9-9e1f-89607d281185-etc-machine-id\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015731 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-ceph\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015750 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-ovsdbserver-sb\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015780 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015801 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015831 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-openstack-edpm-ipam\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015849 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015870 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015896 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data-custom\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015913 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7293dea-c387-4da9-9e1f-89607d281185-logs\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015929 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8jw9\" (UniqueName: \"kubernetes.io/projected/8d6b76db-f778-4b7e-9dab-e0fa5921458e-kube-api-access-h8jw9\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015947 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rzhn\" (UniqueName: \"kubernetes.io/projected/b7293dea-c387-4da9-9e1f-89607d281185-kube-api-access-5rzhn\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015975 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.015992 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-scripts\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.016578 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-config\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.016647 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.016716 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-dns-svc\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.017211 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-ovsdbserver-nb\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.017268 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-openstack-edpm-ipam\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.017381 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.020840 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d6b76db-f778-4b7e-9dab-e0fa5921458e-ovsdbserver-sb\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.022035 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.022358 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.022821 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-scripts\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.029912 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.032769 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-ceph\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.033456 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzqd9\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-kube-api-access-bzqd9\") pod \"manila-share-share1-0\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.035954 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8jw9\" (UniqueName: \"kubernetes.io/projected/8d6b76db-f778-4b7e-9dab-e0fa5921458e-kube-api-access-h8jw9\") pod \"dnsmasq-dns-78f48d6b7c-zbdgc\" (UID: \"8d6b76db-f778-4b7e-9dab-e0fa5921458e\") " pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.048951 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.066968 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.117983 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b7293dea-c387-4da9-9e1f-89607d281185-etc-machine-id\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.118052 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.118086 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.118116 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data-custom\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.118133 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7293dea-c387-4da9-9e1f-89607d281185-logs\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.118155 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rzhn\" (UniqueName: \"kubernetes.io/projected/b7293dea-c387-4da9-9e1f-89607d281185-kube-api-access-5rzhn\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.118181 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-scripts\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.123442 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b7293dea-c387-4da9-9e1f-89607d281185-etc-machine-id\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.123811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7293dea-c387-4da9-9e1f-89607d281185-logs\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.127585 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data-custom\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.132754 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-scripts\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.137452 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.141863 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.141908 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rzhn\" (UniqueName: \"kubernetes.io/projected/b7293dea-c387-4da9-9e1f-89607d281185-kube-api-access-5rzhn\") pod \"manila-api-0\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.230665 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.234869 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.234915 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.234951 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.235651 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cbcdbc5dae82463d722971b6c0e1597afa8425f378dbbe8431940fc4c08a86af"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.235695 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://cbcdbc5dae82463d722971b6c0e1597afa8425f378dbbe8431940fc4c08a86af" gracePeriod=600 Nov 24 20:11:45 crc kubenswrapper[5035]: I1124 20:11:45.531794 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:45.669013 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78f48d6b7c-zbdgc"] Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:45.788271 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:45.888218 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.241512 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b7293dea-c387-4da9-9e1f-89607d281185","Type":"ContainerStarted","Data":"26ba2458563493e86287a0358cde4a71022685ff5752b351947ce96d1d2b8ff3"} Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.242696 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5384a47d-227f-4f78-a66a-f1ca86068978","Type":"ContainerStarted","Data":"1f1d90b734bc885dc96cd6b1c468a107c2b5609f1a08e8fb43ac5841764e5f91"} Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.244685 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="cbcdbc5dae82463d722971b6c0e1597afa8425f378dbbe8431940fc4c08a86af" exitCode=0 Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.244732 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"cbcdbc5dae82463d722971b6c0e1597afa8425f378dbbe8431940fc4c08a86af"} Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.244754 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1"} Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.244794 5035 scope.go:117] "RemoveContainer" containerID="b56852782bd40a40fabaf006315808570f07aa9cddce5aa663d792f295a9f19e" Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.252104 5035 generic.go:334] "Generic (PLEG): container finished" podID="8d6b76db-f778-4b7e-9dab-e0fa5921458e" containerID="3a3032ef0a52193e9ff4627acaa55db0e3553fe05ad40ef2a8483c785f74e37a" exitCode=0 Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.252197 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" event={"ID":"8d6b76db-f778-4b7e-9dab-e0fa5921458e","Type":"ContainerDied","Data":"3a3032ef0a52193e9ff4627acaa55db0e3553fe05ad40ef2a8483c785f74e37a"} Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.252231 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" event={"ID":"8d6b76db-f778-4b7e-9dab-e0fa5921458e","Type":"ContainerStarted","Data":"ff66b4caef4a9354cd736af8ae7b9c7c260bf3bb75f082640c25817ea62ed796"} Nov 24 20:11:46 crc kubenswrapper[5035]: I1124 20:11:46.255232 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f2b24f0f-8067-46eb-bf1c-b2040a20ebda","Type":"ContainerStarted","Data":"29911a60773665ecfaa490c538bae646d9154a85e038c815e8e1f6377d1ce062"} Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.277695 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b7293dea-c387-4da9-9e1f-89607d281185","Type":"ContainerStarted","Data":"8cbb8a0e51ae58b85b0b3e968ac091ed2c80a24e8c08ce993c982273e302e14b"} Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.278262 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.278306 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b7293dea-c387-4da9-9e1f-89607d281185","Type":"ContainerStarted","Data":"1b2f1fd0fc49f60fd5335f5860e55d3330cc014e123e425555a7655098a19e75"} Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.308878 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.308860066 podStartE2EDuration="3.308860066s" podCreationTimestamp="2025-11-24 20:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:11:47.301394285 +0000 UTC m=+3445.823900552" watchObservedRunningTime="2025-11-24 20:11:47.308860066 +0000 UTC m=+3445.831366323" Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.327361 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" event={"ID":"8d6b76db-f778-4b7e-9dab-e0fa5921458e","Type":"ContainerStarted","Data":"43666418fc4d8123f59a2bfb39d4494da08b6247c960da07493f24439e983470"} Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.327457 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.354271 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f2b24f0f-8067-46eb-bf1c-b2040a20ebda","Type":"ContainerStarted","Data":"0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04"} Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.357669 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" podStartSLOduration=3.3576496430000002 podStartE2EDuration="3.357649643s" podCreationTimestamp="2025-11-24 20:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:11:47.352043992 +0000 UTC m=+3445.874550249" watchObservedRunningTime="2025-11-24 20:11:47.357649643 +0000 UTC m=+3445.880155900" Nov 24 20:11:47 crc kubenswrapper[5035]: I1124 20:11:47.586851 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:48 crc kubenswrapper[5035]: I1124 20:11:48.368821 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f2b24f0f-8067-46eb-bf1c-b2040a20ebda","Type":"ContainerStarted","Data":"fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3"} Nov 24 20:11:48 crc kubenswrapper[5035]: I1124 20:11:48.388716 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.430232372 podStartE2EDuration="4.38870007s" podCreationTimestamp="2025-11-24 20:11:44 +0000 UTC" firstStartedPulling="2025-11-24 20:11:45.621377594 +0000 UTC m=+3444.143883851" lastFinishedPulling="2025-11-24 20:11:46.579845292 +0000 UTC m=+3445.102351549" observedRunningTime="2025-11-24 20:11:48.384668111 +0000 UTC m=+3446.907174368" watchObservedRunningTime="2025-11-24 20:11:48.38870007 +0000 UTC m=+3446.911206327" Nov 24 20:11:49 crc kubenswrapper[5035]: I1124 20:11:49.375842 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api-log" containerID="cri-o://1b2f1fd0fc49f60fd5335f5860e55d3330cc014e123e425555a7655098a19e75" gracePeriod=30 Nov 24 20:11:49 crc kubenswrapper[5035]: I1124 20:11:49.375886 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api" containerID="cri-o://8cbb8a0e51ae58b85b0b3e968ac091ed2c80a24e8c08ce993c982273e302e14b" gracePeriod=30 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.173637 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.174501 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-central-agent" containerID="cri-o://38101215349c39cf47ca984ecf78609bd7b92ee341d3a7a8bf98ddee7cb3d1bf" gracePeriod=30 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.174582 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-notification-agent" containerID="cri-o://16643cedcf8ef7af098ed9145fb085b98d72a5c98bf66401e0377161fcd54950" gracePeriod=30 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.174576 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="sg-core" containerID="cri-o://335c8dcab2716763f8973d61075470d1db17488e9d6e11a991abe8e82e7cf7b2" gracePeriod=30 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.174577 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="proxy-httpd" containerID="cri-o://e4c3884585d3e48016841fb56e521726b56a30dba594469d70c110ca1cefe0c3" gracePeriod=30 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.392873 5035 generic.go:334] "Generic (PLEG): container finished" podID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerID="e4c3884585d3e48016841fb56e521726b56a30dba594469d70c110ca1cefe0c3" exitCode=0 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.392934 5035 generic.go:334] "Generic (PLEG): container finished" podID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerID="335c8dcab2716763f8973d61075470d1db17488e9d6e11a991abe8e82e7cf7b2" exitCode=2 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.392959 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerDied","Data":"e4c3884585d3e48016841fb56e521726b56a30dba594469d70c110ca1cefe0c3"} Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.393031 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerDied","Data":"335c8dcab2716763f8973d61075470d1db17488e9d6e11a991abe8e82e7cf7b2"} Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.395102 5035 generic.go:334] "Generic (PLEG): container finished" podID="b7293dea-c387-4da9-9e1f-89607d281185" containerID="8cbb8a0e51ae58b85b0b3e968ac091ed2c80a24e8c08ce993c982273e302e14b" exitCode=0 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.395131 5035 generic.go:334] "Generic (PLEG): container finished" podID="b7293dea-c387-4da9-9e1f-89607d281185" containerID="1b2f1fd0fc49f60fd5335f5860e55d3330cc014e123e425555a7655098a19e75" exitCode=143 Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.395150 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b7293dea-c387-4da9-9e1f-89607d281185","Type":"ContainerDied","Data":"8cbb8a0e51ae58b85b0b3e968ac091ed2c80a24e8c08ce993c982273e302e14b"} Nov 24 20:11:50 crc kubenswrapper[5035]: I1124 20:11:50.395166 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b7293dea-c387-4da9-9e1f-89607d281185","Type":"ContainerDied","Data":"1b2f1fd0fc49f60fd5335f5860e55d3330cc014e123e425555a7655098a19e75"} Nov 24 20:11:51 crc kubenswrapper[5035]: I1124 20:11:51.405320 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerDied","Data":"16643cedcf8ef7af098ed9145fb085b98d72a5c98bf66401e0377161fcd54950"} Nov 24 20:11:51 crc kubenswrapper[5035]: I1124 20:11:51.405329 5035 generic.go:334] "Generic (PLEG): container finished" podID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerID="16643cedcf8ef7af098ed9145fb085b98d72a5c98bf66401e0377161fcd54950" exitCode=0 Nov 24 20:11:51 crc kubenswrapper[5035]: I1124 20:11:51.405598 5035 generic.go:334] "Generic (PLEG): container finished" podID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerID="38101215349c39cf47ca984ecf78609bd7b92ee341d3a7a8bf98ddee7cb3d1bf" exitCode=0 Nov 24 20:11:51 crc kubenswrapper[5035]: I1124 20:11:51.405617 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerDied","Data":"38101215349c39cf47ca984ecf78609bd7b92ee341d3a7a8bf98ddee7cb3d1bf"} Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.752573 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902317 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b7293dea-c387-4da9-9e1f-89607d281185-etc-machine-id\") pod \"b7293dea-c387-4da9-9e1f-89607d281185\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902379 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data\") pod \"b7293dea-c387-4da9-9e1f-89607d281185\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902524 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7293dea-c387-4da9-9e1f-89607d281185-logs\") pod \"b7293dea-c387-4da9-9e1f-89607d281185\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902550 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rzhn\" (UniqueName: \"kubernetes.io/projected/b7293dea-c387-4da9-9e1f-89607d281185-kube-api-access-5rzhn\") pod \"b7293dea-c387-4da9-9e1f-89607d281185\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902627 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data-custom\") pod \"b7293dea-c387-4da9-9e1f-89607d281185\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902675 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-scripts\") pod \"b7293dea-c387-4da9-9e1f-89607d281185\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902688 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7293dea-c387-4da9-9e1f-89607d281185-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b7293dea-c387-4da9-9e1f-89607d281185" (UID: "b7293dea-c387-4da9-9e1f-89607d281185"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.902737 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-combined-ca-bundle\") pod \"b7293dea-c387-4da9-9e1f-89607d281185\" (UID: \"b7293dea-c387-4da9-9e1f-89607d281185\") " Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.903841 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b7293dea-c387-4da9-9e1f-89607d281185-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.904225 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7293dea-c387-4da9-9e1f-89607d281185-logs" (OuterVolumeSpecName: "logs") pod "b7293dea-c387-4da9-9e1f-89607d281185" (UID: "b7293dea-c387-4da9-9e1f-89607d281185"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.916491 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7293dea-c387-4da9-9e1f-89607d281185-kube-api-access-5rzhn" (OuterVolumeSpecName: "kube-api-access-5rzhn") pod "b7293dea-c387-4da9-9e1f-89607d281185" (UID: "b7293dea-c387-4da9-9e1f-89607d281185"). InnerVolumeSpecName "kube-api-access-5rzhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.916579 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b7293dea-c387-4da9-9e1f-89607d281185" (UID: "b7293dea-c387-4da9-9e1f-89607d281185"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.918560 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-scripts" (OuterVolumeSpecName: "scripts") pod "b7293dea-c387-4da9-9e1f-89607d281185" (UID: "b7293dea-c387-4da9-9e1f-89607d281185"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:52 crc kubenswrapper[5035]: I1124 20:11:52.961451 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7293dea-c387-4da9-9e1f-89607d281185" (UID: "b7293dea-c387-4da9-9e1f-89607d281185"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.001227 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.005666 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7293dea-c387-4da9-9e1f-89607d281185-logs\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.005786 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rzhn\" (UniqueName: \"kubernetes.io/projected/b7293dea-c387-4da9-9e1f-89607d281185-kube-api-access-5rzhn\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.005868 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.005921 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.005977 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.087470 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data" (OuterVolumeSpecName: "config-data") pod "b7293dea-c387-4da9-9e1f-89607d281185" (UID: "b7293dea-c387-4da9-9e1f-89607d281185"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.106895 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-run-httpd\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107016 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-scripts\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107079 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-ceilometer-tls-certs\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107146 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-sg-core-conf-yaml\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107210 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-config-data\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107233 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgrbt\" (UniqueName: \"kubernetes.io/projected/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-kube-api-access-bgrbt\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107330 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-combined-ca-bundle\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107399 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-log-httpd\") pod \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\" (UID: \"8d9647e3-bf7f-40db-86c8-299b2b08d5dd\") " Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.107893 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7293dea-c387-4da9-9e1f-89607d281185-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.108446 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.108588 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.120825 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-kube-api-access-bgrbt" (OuterVolumeSpecName: "kube-api-access-bgrbt") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "kube-api-access-bgrbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.136016 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-scripts" (OuterVolumeSpecName: "scripts") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.170583 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.210822 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.210862 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgrbt\" (UniqueName: \"kubernetes.io/projected/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-kube-api-access-bgrbt\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.210880 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.210889 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.210898 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.242561 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.257339 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.265515 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-config-data" (OuterVolumeSpecName: "config-data") pod "8d9647e3-bf7f-40db-86c8-299b2b08d5dd" (UID: "8d9647e3-bf7f-40db-86c8-299b2b08d5dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.314427 5035 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.314458 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.314480 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d9647e3-bf7f-40db-86c8-299b2b08d5dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.429073 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"b7293dea-c387-4da9-9e1f-89607d281185","Type":"ContainerDied","Data":"26ba2458563493e86287a0358cde4a71022685ff5752b351947ce96d1d2b8ff3"} Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.429153 5035 scope.go:117] "RemoveContainer" containerID="8cbb8a0e51ae58b85b0b3e968ac091ed2c80a24e8c08ce993c982273e302e14b" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.429107 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.430777 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5384a47d-227f-4f78-a66a-f1ca86068978","Type":"ContainerStarted","Data":"b3c154c484a02c79c31a9b661f94344a1bac6d86d124cf810998708d53664246"} Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.438130 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d9647e3-bf7f-40db-86c8-299b2b08d5dd","Type":"ContainerDied","Data":"a81d0ac1531422683a37f5eb4d01ea0beac26e0e3deb60170dbca0dacc9cd16a"} Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.438219 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.460997 5035 scope.go:117] "RemoveContainer" containerID="1b2f1fd0fc49f60fd5335f5860e55d3330cc014e123e425555a7655098a19e75" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.464840 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.496777 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.497415 5035 scope.go:117] "RemoveContainer" containerID="e4c3884585d3e48016841fb56e521726b56a30dba594469d70c110ca1cefe0c3" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.511802 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.521099 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527041 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: E1124 20:11:53.527530 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527562 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api" Nov 24 20:11:53 crc kubenswrapper[5035]: E1124 20:11:53.527605 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="sg-core" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527615 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="sg-core" Nov 24 20:11:53 crc kubenswrapper[5035]: E1124 20:11:53.527630 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="proxy-httpd" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527637 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="proxy-httpd" Nov 24 20:11:53 crc kubenswrapper[5035]: E1124 20:11:53.527646 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api-log" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527652 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api-log" Nov 24 20:11:53 crc kubenswrapper[5035]: E1124 20:11:53.527659 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-central-agent" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527665 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-central-agent" Nov 24 20:11:53 crc kubenswrapper[5035]: E1124 20:11:53.527680 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-notification-agent" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527686 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-notification-agent" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527854 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="sg-core" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527863 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="proxy-httpd" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527876 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api-log" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527890 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7293dea-c387-4da9-9e1f-89607d281185" containerName="manila-api" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527896 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-notification-agent" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.527910 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" containerName="ceilometer-central-agent" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.528604 5035 scope.go:117] "RemoveContainer" containerID="335c8dcab2716763f8973d61075470d1db17488e9d6e11a991abe8e82e7cf7b2" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.529456 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.531831 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.532045 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.532175 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.535092 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.537205 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.539921 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.540063 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.540162 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.542304 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.556777 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619689 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-public-tls-certs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-config-data\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619792 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c994550c-82c0-4e4a-9587-0493a16703e7-etc-machine-id\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619819 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqdz9\" (UniqueName: \"kubernetes.io/projected/c994550c-82c0-4e4a-9587-0493a16703e7-kube-api-access-fqdz9\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619834 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-scripts\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619855 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-internal-tls-certs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619899 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619915 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c994550c-82c0-4e4a-9587-0493a16703e7-logs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.619939 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-config-data-custom\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.662892 5035 scope.go:117] "RemoveContainer" containerID="16643cedcf8ef7af098ed9145fb085b98d72a5c98bf66401e0377161fcd54950" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.700573 5035 scope.go:117] "RemoveContainer" containerID="38101215349c39cf47ca984ecf78609bd7b92ee341d3a7a8bf98ddee7cb3d1bf" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.722057 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.722417 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.722560 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-scripts\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.722687 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-public-tls-certs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.722815 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-log-httpd\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.722946 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-config-data\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.723063 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-run-httpd\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.723156 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c994550c-82c0-4e4a-9587-0493a16703e7-etc-machine-id\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.723238 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twplb\" (UniqueName: \"kubernetes.io/projected/8f2f34f7-da42-41ba-89d7-22d1e6423cab-kube-api-access-twplb\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.723337 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqdz9\" (UniqueName: \"kubernetes.io/projected/c994550c-82c0-4e4a-9587-0493a16703e7-kube-api-access-fqdz9\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.723536 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-scripts\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.723929 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-config-data\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.724127 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-internal-tls-certs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.724529 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.724665 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.724788 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c994550c-82c0-4e4a-9587-0493a16703e7-logs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.725842 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-config-data-custom\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.723510 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c994550c-82c0-4e4a-9587-0493a16703e7-etc-machine-id\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.725799 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c994550c-82c0-4e4a-9587-0493a16703e7-logs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.727166 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-public-tls-certs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.727393 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-scripts\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.731855 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-internal-tls-certs\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.733073 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-config-data\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.734210 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-config-data-custom\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.743854 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqdz9\" (UniqueName: \"kubernetes.io/projected/c994550c-82c0-4e4a-9587-0493a16703e7-kube-api-access-fqdz9\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.753973 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c994550c-82c0-4e4a-9587-0493a16703e7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"c994550c-82c0-4e4a-9587-0493a16703e7\") " pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.827904 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-log-httpd\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.827985 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-run-httpd\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.828012 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twplb\" (UniqueName: \"kubernetes.io/projected/8f2f34f7-da42-41ba-89d7-22d1e6423cab-kube-api-access-twplb\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.828051 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-config-data\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.828106 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.828143 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.828159 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.828197 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-scripts\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.829078 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-run-httpd\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.829338 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-log-httpd\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.832036 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-scripts\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.832780 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.834006 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.834685 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.834815 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-config-data\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.849204 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twplb\" (UniqueName: \"kubernetes.io/projected/8f2f34f7-da42-41ba-89d7-22d1e6423cab-kube-api-access-twplb\") pod \"ceilometer-0\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " pod="openstack/ceilometer-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.947141 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 20:11:53 crc kubenswrapper[5035]: I1124 20:11:53.963619 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:11:54 crc kubenswrapper[5035]: I1124 20:11:54.211173 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d9647e3-bf7f-40db-86c8-299b2b08d5dd" path="/var/lib/kubelet/pods/8d9647e3-bf7f-40db-86c8-299b2b08d5dd/volumes" Nov 24 20:11:54 crc kubenswrapper[5035]: I1124 20:11:54.213877 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7293dea-c387-4da9-9e1f-89607d281185" path="/var/lib/kubelet/pods/b7293dea-c387-4da9-9e1f-89607d281185/volumes" Nov 24 20:11:54 crc kubenswrapper[5035]: I1124 20:11:54.478147 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5384a47d-227f-4f78-a66a-f1ca86068978","Type":"ContainerStarted","Data":"5adebacf6c11e242878dac2f91e8630f27a4a20b724f450e4dac240cca853517"} Nov 24 20:11:54 crc kubenswrapper[5035]: I1124 20:11:54.515360 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.673714693 podStartE2EDuration="10.515338306s" podCreationTimestamp="2025-11-24 20:11:44 +0000 UTC" firstStartedPulling="2025-11-24 20:11:45.804396794 +0000 UTC m=+3444.326903051" lastFinishedPulling="2025-11-24 20:11:52.646020397 +0000 UTC m=+3451.168526664" observedRunningTime="2025-11-24 20:11:54.505277565 +0000 UTC m=+3453.027783822" watchObservedRunningTime="2025-11-24 20:11:54.515338306 +0000 UTC m=+3453.037844563" Nov 24 20:11:54 crc kubenswrapper[5035]: I1124 20:11:54.538937 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:11:54 crc kubenswrapper[5035]: I1124 20:11:54.559522 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 20:11:54 crc kubenswrapper[5035]: I1124 20:11:54.933359 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.049645 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.069269 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78f48d6b7c-zbdgc" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.143803 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c58867b6c-btg8h"] Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.144006 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" podUID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerName="dnsmasq-dns" containerID="cri-o://d7e1af7a46b48ded1939fb8cf0093593013ea80c00a3b9891de65e1fab8402d7" gracePeriod=10 Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.496398 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerStarted","Data":"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c"} Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.496680 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerStarted","Data":"59e6e044c247db437dce692d5b4f86137ed110ad6eeb2bdbdff08bb4130bc8d5"} Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.498779 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c994550c-82c0-4e4a-9587-0493a16703e7","Type":"ContainerStarted","Data":"bbaeeaea626b87a7455ead4ed9c4f91798ec0a2e18b12e052f2e1cf8eb04b243"} Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.498810 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c994550c-82c0-4e4a-9587-0493a16703e7","Type":"ContainerStarted","Data":"de8dba2a062f5108c181ed41ecefb292b0fa2060be3b1c1cce31db7fba03be03"} Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.501305 5035 generic.go:334] "Generic (PLEG): container finished" podID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerID="d7e1af7a46b48ded1939fb8cf0093593013ea80c00a3b9891de65e1fab8402d7" exitCode=0 Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.501998 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" event={"ID":"268a0d3d-f4b4-49c2-ae77-9ec588442031","Type":"ContainerDied","Data":"d7e1af7a46b48ded1939fb8cf0093593013ea80c00a3b9891de65e1fab8402d7"} Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.641598 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.701924 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-dns-svc\") pod \"268a0d3d-f4b4-49c2-ae77-9ec588442031\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.702032 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67bbz\" (UniqueName: \"kubernetes.io/projected/268a0d3d-f4b4-49c2-ae77-9ec588442031-kube-api-access-67bbz\") pod \"268a0d3d-f4b4-49c2-ae77-9ec588442031\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.702051 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-openstack-edpm-ipam\") pod \"268a0d3d-f4b4-49c2-ae77-9ec588442031\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.702084 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-sb\") pod \"268a0d3d-f4b4-49c2-ae77-9ec588442031\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.702146 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-nb\") pod \"268a0d3d-f4b4-49c2-ae77-9ec588442031\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.702205 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-config\") pod \"268a0d3d-f4b4-49c2-ae77-9ec588442031\" (UID: \"268a0d3d-f4b4-49c2-ae77-9ec588442031\") " Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.713885 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268a0d3d-f4b4-49c2-ae77-9ec588442031-kube-api-access-67bbz" (OuterVolumeSpecName: "kube-api-access-67bbz") pod "268a0d3d-f4b4-49c2-ae77-9ec588442031" (UID: "268a0d3d-f4b4-49c2-ae77-9ec588442031"). InnerVolumeSpecName "kube-api-access-67bbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.750046 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-config" (OuterVolumeSpecName: "config") pod "268a0d3d-f4b4-49c2-ae77-9ec588442031" (UID: "268a0d3d-f4b4-49c2-ae77-9ec588442031"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.752790 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "268a0d3d-f4b4-49c2-ae77-9ec588442031" (UID: "268a0d3d-f4b4-49c2-ae77-9ec588442031"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.768701 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "268a0d3d-f4b4-49c2-ae77-9ec588442031" (UID: "268a0d3d-f4b4-49c2-ae77-9ec588442031"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.786007 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "268a0d3d-f4b4-49c2-ae77-9ec588442031" (UID: "268a0d3d-f4b4-49c2-ae77-9ec588442031"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.806537 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67bbz\" (UniqueName: \"kubernetes.io/projected/268a0d3d-f4b4-49c2-ae77-9ec588442031-kube-api-access-67bbz\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.806572 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.806581 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.806589 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.806600 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-config\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.820780 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "268a0d3d-f4b4-49c2-ae77-9ec588442031" (UID: "268a0d3d-f4b4-49c2-ae77-9ec588442031"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:11:55 crc kubenswrapper[5035]: I1124 20:11:55.907430 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/268a0d3d-f4b4-49c2-ae77-9ec588442031-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.512751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerStarted","Data":"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612"} Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.515215 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"c994550c-82c0-4e4a-9587-0493a16703e7","Type":"ContainerStarted","Data":"cc1e96540ce5b918c65015cd01136a0ce04f22ce33177239066a61f484d241b5"} Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.515351 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.517202 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" event={"ID":"268a0d3d-f4b4-49c2-ae77-9ec588442031","Type":"ContainerDied","Data":"3f224e6f1e1a6a00662e3ad9be2d83de06207dbe22467572d25b88dd5839a201"} Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.517462 5035 scope.go:117] "RemoveContainer" containerID="d7e1af7a46b48ded1939fb8cf0093593013ea80c00a3b9891de65e1fab8402d7" Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.517899 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c58867b6c-btg8h" Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.550667 5035 scope.go:117] "RemoveContainer" containerID="25b5bc190054feab413a22a038a720c7556311acd08fc2c49bae6c89fc6a72a5" Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.557831 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.557813129 podStartE2EDuration="3.557813129s" podCreationTimestamp="2025-11-24 20:11:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:11:56.551930081 +0000 UTC m=+3455.074436328" watchObservedRunningTime="2025-11-24 20:11:56.557813129 +0000 UTC m=+3455.080319376" Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.590456 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c58867b6c-btg8h"] Nov 24 20:11:56 crc kubenswrapper[5035]: I1124 20:11:56.601719 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c58867b6c-btg8h"] Nov 24 20:11:57 crc kubenswrapper[5035]: I1124 20:11:57.540163 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerStarted","Data":"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff"} Nov 24 20:11:58 crc kubenswrapper[5035]: I1124 20:11:58.210263 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="268a0d3d-f4b4-49c2-ae77-9ec588442031" path="/var/lib/kubelet/pods/268a0d3d-f4b4-49c2-ae77-9ec588442031/volumes" Nov 24 20:11:58 crc kubenswrapper[5035]: I1124 20:11:58.932683 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:11:59 crc kubenswrapper[5035]: I1124 20:11:59.557417 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerStarted","Data":"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72"} Nov 24 20:11:59 crc kubenswrapper[5035]: I1124 20:11:59.557724 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 20:11:59 crc kubenswrapper[5035]: I1124 20:11:59.585661 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.82192158 podStartE2EDuration="6.585640946s" podCreationTimestamp="2025-11-24 20:11:53 +0000 UTC" firstStartedPulling="2025-11-24 20:11:54.574494274 +0000 UTC m=+3453.097000531" lastFinishedPulling="2025-11-24 20:11:58.33821364 +0000 UTC m=+3456.860719897" observedRunningTime="2025-11-24 20:11:59.580443385 +0000 UTC m=+3458.102949642" watchObservedRunningTime="2025-11-24 20:11:59.585640946 +0000 UTC m=+3458.108147213" Nov 24 20:12:00 crc kubenswrapper[5035]: I1124 20:12:00.568916 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="sg-core" containerID="cri-o://95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" gracePeriod=30 Nov 24 20:12:00 crc kubenswrapper[5035]: I1124 20:12:00.568960 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="proxy-httpd" containerID="cri-o://b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" gracePeriod=30 Nov 24 20:12:00 crc kubenswrapper[5035]: I1124 20:12:00.568895 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-central-agent" containerID="cri-o://aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" gracePeriod=30 Nov 24 20:12:00 crc kubenswrapper[5035]: I1124 20:12:00.568969 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-notification-agent" containerID="cri-o://24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" gracePeriod=30 Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.394555 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442119 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-config-data\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442209 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-combined-ca-bundle\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442257 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twplb\" (UniqueName: \"kubernetes.io/projected/8f2f34f7-da42-41ba-89d7-22d1e6423cab-kube-api-access-twplb\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442307 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-sg-core-conf-yaml\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442431 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-ceilometer-tls-certs\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442464 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-run-httpd\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442513 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-log-httpd\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442559 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-scripts\") pod \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\" (UID: \"8f2f34f7-da42-41ba-89d7-22d1e6423cab\") " Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.442912 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.443471 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.447273 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2f34f7-da42-41ba-89d7-22d1e6423cab-kube-api-access-twplb" (OuterVolumeSpecName: "kube-api-access-twplb") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "kube-api-access-twplb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.449433 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-scripts" (OuterVolumeSpecName: "scripts") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.469187 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.491530 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.518129 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.543019 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-config-data" (OuterVolumeSpecName: "config-data") pod "8f2f34f7-da42-41ba-89d7-22d1e6423cab" (UID: "8f2f34f7-da42-41ba-89d7-22d1e6423cab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544465 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544487 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544499 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twplb\" (UniqueName: \"kubernetes.io/projected/8f2f34f7-da42-41ba-89d7-22d1e6423cab-kube-api-access-twplb\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544508 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544516 5035 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544525 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544533 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8f2f34f7-da42-41ba-89d7-22d1e6423cab-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.544540 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f34f7-da42-41ba-89d7-22d1e6423cab-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596361 5035 generic.go:334] "Generic (PLEG): container finished" podID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerID="b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" exitCode=0 Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596392 5035 generic.go:334] "Generic (PLEG): container finished" podID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerID="95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" exitCode=2 Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596400 5035 generic.go:334] "Generic (PLEG): container finished" podID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerID="24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" exitCode=0 Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596411 5035 generic.go:334] "Generic (PLEG): container finished" podID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerID="aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" exitCode=0 Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596433 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerDied","Data":"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72"} Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596462 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerDied","Data":"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff"} Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596473 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerDied","Data":"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612"} Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596484 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerDied","Data":"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c"} Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596496 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8f2f34f7-da42-41ba-89d7-22d1e6423cab","Type":"ContainerDied","Data":"59e6e044c247db437dce692d5b4f86137ed110ad6eeb2bdbdff08bb4130bc8d5"} Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596513 5035 scope.go:117] "RemoveContainer" containerID="b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.596664 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.634278 5035 scope.go:117] "RemoveContainer" containerID="95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.646574 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.658686 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.658786 5035 scope.go:117] "RemoveContainer" containerID="24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.668698 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.669169 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-central-agent" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.669234 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-central-agent" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.669310 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="sg-core" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.669358 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="sg-core" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.669422 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerName="init" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.669468 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerName="init" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.669529 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerName="dnsmasq-dns" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.669574 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerName="dnsmasq-dns" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.669623 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-notification-agent" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.669667 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-notification-agent" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.669740 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="proxy-httpd" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.669787 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="proxy-httpd" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.669984 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="sg-core" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.670054 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="268a0d3d-f4b4-49c2-ae77-9ec588442031" containerName="dnsmasq-dns" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.670104 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="proxy-httpd" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.670158 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-notification-agent" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.670209 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" containerName="ceilometer-central-agent" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.671765 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.674321 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.674517 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.681736 5035 scope.go:117] "RemoveContainer" containerID="aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.682115 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.690734 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748256 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9d5829b-2574-4340-aa87-c7d78011d378-log-httpd\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748401 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-config-data\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748455 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748572 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-scripts\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748607 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748709 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9d5829b-2574-4340-aa87-c7d78011d378-run-httpd\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.748771 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jl4c\" (UniqueName: \"kubernetes.io/projected/f9d5829b-2574-4340-aa87-c7d78011d378-kube-api-access-2jl4c\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.765394 5035 scope.go:117] "RemoveContainer" containerID="b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.766371 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": container with ID starting with b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72 not found: ID does not exist" containerID="b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.766510 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72"} err="failed to get container status \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": rpc error: code = NotFound desc = could not find container \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": container with ID starting with b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.766889 5035 scope.go:117] "RemoveContainer" containerID="95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.767285 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": container with ID starting with 95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff not found: ID does not exist" containerID="95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.767324 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff"} err="failed to get container status \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": rpc error: code = NotFound desc = could not find container \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": container with ID starting with 95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.767346 5035 scope.go:117] "RemoveContainer" containerID="24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.768120 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": container with ID starting with 24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612 not found: ID does not exist" containerID="24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.768154 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612"} err="failed to get container status \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": rpc error: code = NotFound desc = could not find container \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": container with ID starting with 24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.768174 5035 scope.go:117] "RemoveContainer" containerID="aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" Nov 24 20:12:01 crc kubenswrapper[5035]: E1124 20:12:01.768648 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": container with ID starting with aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c not found: ID does not exist" containerID="aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.768674 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c"} err="failed to get container status \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": rpc error: code = NotFound desc = could not find container \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": container with ID starting with aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.768692 5035 scope.go:117] "RemoveContainer" containerID="b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.769006 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72"} err="failed to get container status \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": rpc error: code = NotFound desc = could not find container \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": container with ID starting with b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.769111 5035 scope.go:117] "RemoveContainer" containerID="95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.769461 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff"} err="failed to get container status \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": rpc error: code = NotFound desc = could not find container \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": container with ID starting with 95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.769482 5035 scope.go:117] "RemoveContainer" containerID="24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.769716 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612"} err="failed to get container status \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": rpc error: code = NotFound desc = could not find container \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": container with ID starting with 24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.769734 5035 scope.go:117] "RemoveContainer" containerID="aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.769983 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c"} err="failed to get container status \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": rpc error: code = NotFound desc = could not find container \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": container with ID starting with aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.770064 5035 scope.go:117] "RemoveContainer" containerID="b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.770382 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72"} err="failed to get container status \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": rpc error: code = NotFound desc = could not find container \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": container with ID starting with b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.770407 5035 scope.go:117] "RemoveContainer" containerID="95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.770687 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff"} err="failed to get container status \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": rpc error: code = NotFound desc = could not find container \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": container with ID starting with 95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.770707 5035 scope.go:117] "RemoveContainer" containerID="24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.770983 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612"} err="failed to get container status \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": rpc error: code = NotFound desc = could not find container \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": container with ID starting with 24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.771009 5035 scope.go:117] "RemoveContainer" containerID="aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.771323 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c"} err="failed to get container status \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": rpc error: code = NotFound desc = could not find container \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": container with ID starting with aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.771348 5035 scope.go:117] "RemoveContainer" containerID="b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.771742 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72"} err="failed to get container status \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": rpc error: code = NotFound desc = could not find container \"b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72\": container with ID starting with b57535bddcb594addb043ef885ee01590610d5470cccb419a0cbb572a8400a72 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.771785 5035 scope.go:117] "RemoveContainer" containerID="95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.772098 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff"} err="failed to get container status \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": rpc error: code = NotFound desc = could not find container \"95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff\": container with ID starting with 95f178bea3f29cca3eb1331d2176d72c1c7a964f3bd87996c0737fa36e61bcff not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.772121 5035 scope.go:117] "RemoveContainer" containerID="24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.772477 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612"} err="failed to get container status \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": rpc error: code = NotFound desc = could not find container \"24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612\": container with ID starting with 24672145b09d970bb736e2933e4fea0cb6285f00c0fe6fca57770491c9060612 not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.772503 5035 scope.go:117] "RemoveContainer" containerID="aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.772834 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c"} err="failed to get container status \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": rpc error: code = NotFound desc = could not find container \"aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c\": container with ID starting with aa141fa93f903c37c9b2a2d6cb579aeae3d5889eac3c4ee2af189fcc4824728c not found: ID does not exist" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850623 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jl4c\" (UniqueName: \"kubernetes.io/projected/f9d5829b-2574-4340-aa87-c7d78011d378-kube-api-access-2jl4c\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850684 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9d5829b-2574-4340-aa87-c7d78011d378-log-httpd\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850741 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-config-data\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850778 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850808 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-scripts\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850824 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850850 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.850872 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9d5829b-2574-4340-aa87-c7d78011d378-run-httpd\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.851390 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9d5829b-2574-4340-aa87-c7d78011d378-run-httpd\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.852015 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9d5829b-2574-4340-aa87-c7d78011d378-log-httpd\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.855368 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.855469 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-scripts\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.856274 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.856918 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-config-data\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.857277 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d5829b-2574-4340-aa87-c7d78011d378-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:01 crc kubenswrapper[5035]: I1124 20:12:01.870492 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jl4c\" (UniqueName: \"kubernetes.io/projected/f9d5829b-2574-4340-aa87-c7d78011d378-kube-api-access-2jl4c\") pod \"ceilometer-0\" (UID: \"f9d5829b-2574-4340-aa87-c7d78011d378\") " pod="openstack/ceilometer-0" Nov 24 20:12:02 crc kubenswrapper[5035]: I1124 20:12:02.055126 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 20:12:02 crc kubenswrapper[5035]: I1124 20:12:02.238670 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2f34f7-da42-41ba-89d7-22d1e6423cab" path="/var/lib/kubelet/pods/8f2f34f7-da42-41ba-89d7-22d1e6423cab/volumes" Nov 24 20:12:02 crc kubenswrapper[5035]: I1124 20:12:02.569439 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 20:12:02 crc kubenswrapper[5035]: I1124 20:12:02.605011 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9d5829b-2574-4340-aa87-c7d78011d378","Type":"ContainerStarted","Data":"e977635668053b94503735dccb98f7e8758de8cd2159a0591b63966581c54708"} Nov 24 20:12:03 crc kubenswrapper[5035]: I1124 20:12:03.614848 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9d5829b-2574-4340-aa87-c7d78011d378","Type":"ContainerStarted","Data":"70934bfccfbe579815267885c43ef235edb3c569eb2835682d81ce37f5848f61"} Nov 24 20:12:04 crc kubenswrapper[5035]: I1124 20:12:04.628519 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9d5829b-2574-4340-aa87-c7d78011d378","Type":"ContainerStarted","Data":"34ea6d3862c01a8342f3145620986a08698f392283d6c9bcd28adeaed9107115"} Nov 24 20:12:04 crc kubenswrapper[5035]: I1124 20:12:04.628926 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9d5829b-2574-4340-aa87-c7d78011d378","Type":"ContainerStarted","Data":"502bb897a2c7b30c6bcbc481b05af573e7a65708890bb9453718f6f9ecd8d4be"} Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.468900 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.526054 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.647893 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9d5829b-2574-4340-aa87-c7d78011d378","Type":"ContainerStarted","Data":"5d8e9bc40ea6e3460bf748938be95bc55cb9a3747c0922163519b655fb47231f"} Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.647933 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="manila-scheduler" containerID="cri-o://0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04" gracePeriod=30 Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.648062 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="probe" containerID="cri-o://fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3" gracePeriod=30 Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.736138 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.756181 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6442640600000002 podStartE2EDuration="5.756164286s" podCreationTimestamp="2025-11-24 20:12:01 +0000 UTC" firstStartedPulling="2025-11-24 20:12:02.585316031 +0000 UTC m=+3461.107822278" lastFinishedPulling="2025-11-24 20:12:05.697216257 +0000 UTC m=+3464.219722504" observedRunningTime="2025-11-24 20:12:06.676396943 +0000 UTC m=+3465.198903200" watchObservedRunningTime="2025-11-24 20:12:06.756164286 +0000 UTC m=+3465.278670543" Nov 24 20:12:06 crc kubenswrapper[5035]: I1124 20:12:06.795801 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:12:07 crc kubenswrapper[5035]: I1124 20:12:07.657503 5035 generic.go:334] "Generic (PLEG): container finished" podID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerID="fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3" exitCode=0 Nov 24 20:12:07 crc kubenswrapper[5035]: I1124 20:12:07.657538 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f2b24f0f-8067-46eb-bf1c-b2040a20ebda","Type":"ContainerDied","Data":"fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3"} Nov 24 20:12:07 crc kubenswrapper[5035]: I1124 20:12:07.657698 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="manila-share" containerID="cri-o://b3c154c484a02c79c31a9b661f94344a1bac6d86d124cf810998708d53664246" gracePeriod=30 Nov 24 20:12:07 crc kubenswrapper[5035]: I1124 20:12:07.658038 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="probe" containerID="cri-o://5adebacf6c11e242878dac2f91e8630f27a4a20b724f450e4dac240cca853517" gracePeriod=30 Nov 24 20:12:07 crc kubenswrapper[5035]: I1124 20:12:07.658378 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.667029 5035 generic.go:334] "Generic (PLEG): container finished" podID="5384a47d-227f-4f78-a66a-f1ca86068978" containerID="5adebacf6c11e242878dac2f91e8630f27a4a20b724f450e4dac240cca853517" exitCode=0 Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.667464 5035 generic.go:334] "Generic (PLEG): container finished" podID="5384a47d-227f-4f78-a66a-f1ca86068978" containerID="b3c154c484a02c79c31a9b661f94344a1bac6d86d124cf810998708d53664246" exitCode=1 Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.667219 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5384a47d-227f-4f78-a66a-f1ca86068978","Type":"ContainerDied","Data":"5adebacf6c11e242878dac2f91e8630f27a4a20b724f450e4dac240cca853517"} Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.668396 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5384a47d-227f-4f78-a66a-f1ca86068978","Type":"ContainerDied","Data":"b3c154c484a02c79c31a9b661f94344a1bac6d86d124cf810998708d53664246"} Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.668410 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"5384a47d-227f-4f78-a66a-f1ca86068978","Type":"ContainerDied","Data":"1f1d90b734bc885dc96cd6b1c468a107c2b5609f1a08e8fb43ac5841764e5f91"} Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.668421 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f1d90b734bc885dc96cd6b1c468a107c2b5609f1a08e8fb43ac5841764e5f91" Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.870663 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988243 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-combined-ca-bundle\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988480 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-scripts\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988562 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-etc-machine-id\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988644 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988711 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data-custom\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988754 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-ceph\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988793 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-var-lib-manila\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.988840 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzqd9\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-kube-api-access-bzqd9\") pod \"5384a47d-227f-4f78-a66a-f1ca86068978\" (UID: \"5384a47d-227f-4f78-a66a-f1ca86068978\") " Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.991841 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.995590 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.997724 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-kube-api-access-bzqd9" (OuterVolumeSpecName: "kube-api-access-bzqd9") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "kube-api-access-bzqd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:12:08 crc kubenswrapper[5035]: I1124 20:12:08.998955 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-ceph" (OuterVolumeSpecName: "ceph") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.000624 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-scripts" (OuterVolumeSpecName: "scripts") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.005801 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.056993 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.063601 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.091365 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.091571 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.091625 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.091675 5035 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/5384a47d-227f-4f78-a66a-f1ca86068978-var-lib-manila\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.091723 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzqd9\" (UniqueName: \"kubernetes.io/projected/5384a47d-227f-4f78-a66a-f1ca86068978-kube-api-access-bzqd9\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.091809 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.091861 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.139450 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data" (OuterVolumeSpecName: "config-data") pod "5384a47d-227f-4f78-a66a-f1ca86068978" (UID: "5384a47d-227f-4f78-a66a-f1ca86068978"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.200049 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data\") pod \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.200160 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data-custom\") pod \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.200266 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-combined-ca-bundle\") pod \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.200358 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-254q4\" (UniqueName: \"kubernetes.io/projected/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-kube-api-access-254q4\") pod \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.200459 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-scripts\") pod \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.200519 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-etc-machine-id\") pod \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\" (UID: \"f2b24f0f-8067-46eb-bf1c-b2040a20ebda\") " Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.200972 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5384a47d-227f-4f78-a66a-f1ca86068978-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.201133 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f2b24f0f-8067-46eb-bf1c-b2040a20ebda" (UID: "f2b24f0f-8067-46eb-bf1c-b2040a20ebda"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.204368 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f2b24f0f-8067-46eb-bf1c-b2040a20ebda" (UID: "f2b24f0f-8067-46eb-bf1c-b2040a20ebda"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.208170 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-kube-api-access-254q4" (OuterVolumeSpecName: "kube-api-access-254q4") pod "f2b24f0f-8067-46eb-bf1c-b2040a20ebda" (UID: "f2b24f0f-8067-46eb-bf1c-b2040a20ebda"). InnerVolumeSpecName "kube-api-access-254q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.210548 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-scripts" (OuterVolumeSpecName: "scripts") pod "f2b24f0f-8067-46eb-bf1c-b2040a20ebda" (UID: "f2b24f0f-8067-46eb-bf1c-b2040a20ebda"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.267560 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2b24f0f-8067-46eb-bf1c-b2040a20ebda" (UID: "f2b24f0f-8067-46eb-bf1c-b2040a20ebda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.290860 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data" (OuterVolumeSpecName: "config-data") pod "f2b24f0f-8067-46eb-bf1c-b2040a20ebda" (UID: "f2b24f0f-8067-46eb-bf1c-b2040a20ebda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.303765 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.303957 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.304039 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-254q4\" (UniqueName: \"kubernetes.io/projected/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-kube-api-access-254q4\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.304096 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.304147 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.304201 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b24f0f-8067-46eb-bf1c-b2040a20ebda-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.716330 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f2b24f0f-8067-46eb-bf1c-b2040a20ebda","Type":"ContainerDied","Data":"0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04"} Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.716390 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.716401 5035 scope.go:117] "RemoveContainer" containerID="fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.716280 5035 generic.go:334] "Generic (PLEG): container finished" podID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerID="0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04" exitCode=0 Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.716641 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f2b24f0f-8067-46eb-bf1c-b2040a20ebda","Type":"ContainerDied","Data":"29911a60773665ecfaa490c538bae646d9154a85e038c815e8e1f6377d1ce062"} Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.716996 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.752527 5035 scope.go:117] "RemoveContainer" containerID="0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.778744 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.791982 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805055 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: E1124 20:12:09.805562 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="manila-share" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805582 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="manila-share" Nov 24 20:12:09 crc kubenswrapper[5035]: E1124 20:12:09.805608 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="probe" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805615 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="probe" Nov 24 20:12:09 crc kubenswrapper[5035]: E1124 20:12:09.805623 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="manila-scheduler" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805630 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="manila-scheduler" Nov 24 20:12:09 crc kubenswrapper[5035]: E1124 20:12:09.805637 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="probe" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805643 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="probe" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805810 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="probe" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805832 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="probe" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805845 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" containerName="manila-scheduler" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.805859 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" containerName="manila-share" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.811409 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.813168 5035 scope.go:117] "RemoveContainer" containerID="fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.813466 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 24 20:12:09 crc kubenswrapper[5035]: E1124 20:12:09.813632 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3\": container with ID starting with fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3 not found: ID does not exist" containerID="fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.813659 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3"} err="failed to get container status \"fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3\": rpc error: code = NotFound desc = could not find container \"fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3\": container with ID starting with fb505e5f9f8cd37d7c9b046749394279c562b3ce9fcaa4c0b3055c24679b4ba3 not found: ID does not exist" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.813683 5035 scope.go:117] "RemoveContainer" containerID="0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04" Nov 24 20:12:09 crc kubenswrapper[5035]: E1124 20:12:09.814543 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04\": container with ID starting with 0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04 not found: ID does not exist" containerID="0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.814576 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04"} err="failed to get container status \"0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04\": rpc error: code = NotFound desc = could not find container \"0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04\": container with ID starting with 0215b4878b33883348f77c8227a2a91890e53e5ec2f7081b3a85a4341941fd04 not found: ID does not exist" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.815623 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/3beb7c92-aa3c-425d-9964-1dfa44680669-kube-api-access-59s5w\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.815680 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.815706 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-config-data\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.815730 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.815750 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-scripts\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.815828 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3beb7c92-aa3c-425d-9964-1dfa44680669-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.844847 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.856732 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.863508 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.870058 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.871633 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.874201 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.877072 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917636 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-scripts\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917686 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53e5b6e3-1cf1-4537-8263-553481eb0242-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917734 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3beb7c92-aa3c-425d-9964-1dfa44680669-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917777 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-config-data\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917776 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3beb7c92-aa3c-425d-9964-1dfa44680669-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917806 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917877 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/3beb7c92-aa3c-425d-9964-1dfa44680669-kube-api-access-59s5w\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917918 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/53e5b6e3-1cf1-4537-8263-553481eb0242-ceph\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917939 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917969 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.917990 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kqdl\" (UniqueName: \"kubernetes.io/projected/53e5b6e3-1cf1-4537-8263-553481eb0242-kube-api-access-7kqdl\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.918008 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-config-data\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.918127 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.918187 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-scripts\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.918229 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/53e5b6e3-1cf1-4537-8263-553481eb0242-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.921325 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.921771 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.922799 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-config-data\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.932840 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beb7c92-aa3c-425d-9964-1dfa44680669-scripts\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:09 crc kubenswrapper[5035]: I1124 20:12:09.936033 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59s5w\" (UniqueName: \"kubernetes.io/projected/3beb7c92-aa3c-425d-9964-1dfa44680669-kube-api-access-59s5w\") pod \"manila-scheduler-0\" (UID: \"3beb7c92-aa3c-425d-9964-1dfa44680669\") " pod="openstack/manila-scheduler-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.019200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-config-data\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.019256 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.019318 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/53e5b6e3-1cf1-4537-8263-553481eb0242-ceph\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.019343 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.019394 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kqdl\" (UniqueName: \"kubernetes.io/projected/53e5b6e3-1cf1-4537-8263-553481eb0242-kube-api-access-7kqdl\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.019560 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/53e5b6e3-1cf1-4537-8263-553481eb0242-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.019869 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/53e5b6e3-1cf1-4537-8263-553481eb0242-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.020015 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-scripts\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.020050 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53e5b6e3-1cf1-4537-8263-553481eb0242-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.020115 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53e5b6e3-1cf1-4537-8263-553481eb0242-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.022395 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.022630 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.022886 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-config-data\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.027282 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/53e5b6e3-1cf1-4537-8263-553481eb0242-ceph\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.027376 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e5b6e3-1cf1-4537-8263-553481eb0242-scripts\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.040193 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kqdl\" (UniqueName: \"kubernetes.io/projected/53e5b6e3-1cf1-4537-8263-553481eb0242-kube-api-access-7kqdl\") pod \"manila-share-share1-0\" (UID: \"53e5b6e3-1cf1-4537-8263-553481eb0242\") " pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.130817 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.187749 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.210376 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5384a47d-227f-4f78-a66a-f1ca86068978" path="/var/lib/kubelet/pods/5384a47d-227f-4f78-a66a-f1ca86068978/volumes" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.211278 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2b24f0f-8067-46eb-bf1c-b2040a20ebda" path="/var/lib/kubelet/pods/f2b24f0f-8067-46eb-bf1c-b2040a20ebda/volumes" Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.595765 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.714328 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 20:12:10 crc kubenswrapper[5035]: I1124 20:12:10.729997 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3beb7c92-aa3c-425d-9964-1dfa44680669","Type":"ContainerStarted","Data":"d1782d6e2b5b94ab47cc07922b035a70c2ceafbc5534ae7037af39b093a88e30"} Nov 24 20:12:11 crc kubenswrapper[5035]: I1124 20:12:11.744154 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3beb7c92-aa3c-425d-9964-1dfa44680669","Type":"ContainerStarted","Data":"38ffb250e2cba65658a3f0f9d19041781c429237341c5fe4789d0c835f593720"} Nov 24 20:12:11 crc kubenswrapper[5035]: I1124 20:12:11.746550 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3beb7c92-aa3c-425d-9964-1dfa44680669","Type":"ContainerStarted","Data":"87ec726ece670bd67d8f19d616bf3f9c2b77eb037a4270e2a072cfc580ad7e13"} Nov 24 20:12:11 crc kubenswrapper[5035]: I1124 20:12:11.748958 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"53e5b6e3-1cf1-4537-8263-553481eb0242","Type":"ContainerStarted","Data":"2d9424d4753b53a33be7c3ab7289352008811b66eb473c93931e618586360fd5"} Nov 24 20:12:11 crc kubenswrapper[5035]: I1124 20:12:11.749155 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"53e5b6e3-1cf1-4537-8263-553481eb0242","Type":"ContainerStarted","Data":"2f82ad8082a9b2d7cd7cc371033c8758b7c9d2220481636e8e622fce411eaf9c"} Nov 24 20:12:11 crc kubenswrapper[5035]: I1124 20:12:11.749322 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"53e5b6e3-1cf1-4537-8263-553481eb0242","Type":"ContainerStarted","Data":"e19c925712035bcd3cb1a524888e6f1ff4f0aef721c3b341c00a819302f628f0"} Nov 24 20:12:11 crc kubenswrapper[5035]: I1124 20:12:11.774811 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.774788689 podStartE2EDuration="2.774788689s" podCreationTimestamp="2025-11-24 20:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:12:11.759152667 +0000 UTC m=+3470.281658924" watchObservedRunningTime="2025-11-24 20:12:11.774788689 +0000 UTC m=+3470.297294956" Nov 24 20:12:11 crc kubenswrapper[5035]: I1124 20:12:11.784074 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.784046109 podStartE2EDuration="2.784046109s" podCreationTimestamp="2025-11-24 20:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:12:11.775722055 +0000 UTC m=+3470.298228312" watchObservedRunningTime="2025-11-24 20:12:11.784046109 +0000 UTC m=+3470.306552396" Nov 24 20:12:15 crc kubenswrapper[5035]: I1124 20:12:15.153314 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 24 20:12:20 crc kubenswrapper[5035]: I1124 20:12:20.131912 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 24 20:12:20 crc kubenswrapper[5035]: I1124 20:12:20.188802 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 24 20:12:21 crc kubenswrapper[5035]: I1124 20:12:21.570526 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 24 20:12:31 crc kubenswrapper[5035]: I1124 20:12:31.592248 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 24 20:12:32 crc kubenswrapper[5035]: I1124 20:12:32.062684 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.692079 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m64gl"] Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.695379 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.705816 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m64gl"] Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.840878 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n4pw\" (UniqueName: \"kubernetes.io/projected/ada10e45-c886-41a0-a6b4-2d16debe62a7-kube-api-access-5n4pw\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.841229 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-utilities\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.841465 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-catalog-content\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.943104 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n4pw\" (UniqueName: \"kubernetes.io/projected/ada10e45-c886-41a0-a6b4-2d16debe62a7-kube-api-access-5n4pw\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.943246 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-utilities\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.943355 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-catalog-content\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.943946 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-catalog-content\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.943972 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-utilities\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:41 crc kubenswrapper[5035]: I1124 20:12:41.962495 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n4pw\" (UniqueName: \"kubernetes.io/projected/ada10e45-c886-41a0-a6b4-2d16debe62a7-kube-api-access-5n4pw\") pod \"community-operators-m64gl\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:42 crc kubenswrapper[5035]: I1124 20:12:42.016865 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:42 crc kubenswrapper[5035]: I1124 20:12:42.590346 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m64gl"] Nov 24 20:12:43 crc kubenswrapper[5035]: I1124 20:12:43.114796 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerStarted","Data":"c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c"} Nov 24 20:12:43 crc kubenswrapper[5035]: I1124 20:12:43.115196 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerStarted","Data":"d6707391d7d2d81f550525289faa7766320405d3bd283a375108414e9ed694ba"} Nov 24 20:12:44 crc kubenswrapper[5035]: I1124 20:12:44.125787 5035 generic.go:334] "Generic (PLEG): container finished" podID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerID="c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c" exitCode=0 Nov 24 20:12:44 crc kubenswrapper[5035]: I1124 20:12:44.125850 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerDied","Data":"c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c"} Nov 24 20:12:44 crc kubenswrapper[5035]: I1124 20:12:44.127995 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:12:45 crc kubenswrapper[5035]: I1124 20:12:45.138879 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerStarted","Data":"648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b"} Nov 24 20:12:46 crc kubenswrapper[5035]: I1124 20:12:46.159033 5035 generic.go:334] "Generic (PLEG): container finished" podID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerID="648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b" exitCode=0 Nov 24 20:12:46 crc kubenswrapper[5035]: I1124 20:12:46.159131 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerDied","Data":"648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b"} Nov 24 20:12:47 crc kubenswrapper[5035]: I1124 20:12:47.174063 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerStarted","Data":"52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4"} Nov 24 20:12:47 crc kubenswrapper[5035]: I1124 20:12:47.213722 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m64gl" podStartSLOduration=3.751002081 podStartE2EDuration="6.213698045s" podCreationTimestamp="2025-11-24 20:12:41 +0000 UTC" firstStartedPulling="2025-11-24 20:12:44.127574336 +0000 UTC m=+3502.650080583" lastFinishedPulling="2025-11-24 20:12:46.59027029 +0000 UTC m=+3505.112776547" observedRunningTime="2025-11-24 20:12:47.209786619 +0000 UTC m=+3505.732292876" watchObservedRunningTime="2025-11-24 20:12:47.213698045 +0000 UTC m=+3505.736204312" Nov 24 20:12:52 crc kubenswrapper[5035]: I1124 20:12:52.017928 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:52 crc kubenswrapper[5035]: I1124 20:12:52.018584 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:52 crc kubenswrapper[5035]: I1124 20:12:52.064415 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:52 crc kubenswrapper[5035]: I1124 20:12:52.357705 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:52 crc kubenswrapper[5035]: I1124 20:12:52.423875 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m64gl"] Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.300464 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m64gl" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="registry-server" containerID="cri-o://52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4" gracePeriod=2 Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.814962 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.925312 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-catalog-content\") pod \"ada10e45-c886-41a0-a6b4-2d16debe62a7\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.925730 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n4pw\" (UniqueName: \"kubernetes.io/projected/ada10e45-c886-41a0-a6b4-2d16debe62a7-kube-api-access-5n4pw\") pod \"ada10e45-c886-41a0-a6b4-2d16debe62a7\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.925828 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-utilities\") pod \"ada10e45-c886-41a0-a6b4-2d16debe62a7\" (UID: \"ada10e45-c886-41a0-a6b4-2d16debe62a7\") " Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.926733 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-utilities" (OuterVolumeSpecName: "utilities") pod "ada10e45-c886-41a0-a6b4-2d16debe62a7" (UID: "ada10e45-c886-41a0-a6b4-2d16debe62a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.930916 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ada10e45-c886-41a0-a6b4-2d16debe62a7-kube-api-access-5n4pw" (OuterVolumeSpecName: "kube-api-access-5n4pw") pod "ada10e45-c886-41a0-a6b4-2d16debe62a7" (UID: "ada10e45-c886-41a0-a6b4-2d16debe62a7"). InnerVolumeSpecName "kube-api-access-5n4pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:12:54 crc kubenswrapper[5035]: I1124 20:12:54.965478 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ada10e45-c886-41a0-a6b4-2d16debe62a7" (UID: "ada10e45-c886-41a0-a6b4-2d16debe62a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.027753 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.027783 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n4pw\" (UniqueName: \"kubernetes.io/projected/ada10e45-c886-41a0-a6b4-2d16debe62a7-kube-api-access-5n4pw\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.027794 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ada10e45-c886-41a0-a6b4-2d16debe62a7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.314400 5035 generic.go:334] "Generic (PLEG): container finished" podID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerID="52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4" exitCode=0 Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.314461 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerDied","Data":"52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4"} Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.314499 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m64gl" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.314523 5035 scope.go:117] "RemoveContainer" containerID="52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.314506 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m64gl" event={"ID":"ada10e45-c886-41a0-a6b4-2d16debe62a7","Type":"ContainerDied","Data":"d6707391d7d2d81f550525289faa7766320405d3bd283a375108414e9ed694ba"} Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.355433 5035 scope.go:117] "RemoveContainer" containerID="648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.362697 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m64gl"] Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.369819 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m64gl"] Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.392048 5035 scope.go:117] "RemoveContainer" containerID="c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.456752 5035 scope.go:117] "RemoveContainer" containerID="52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4" Nov 24 20:12:55 crc kubenswrapper[5035]: E1124 20:12:55.457898 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4\": container with ID starting with 52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4 not found: ID does not exist" containerID="52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.457990 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4"} err="failed to get container status \"52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4\": rpc error: code = NotFound desc = could not find container \"52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4\": container with ID starting with 52825791470e5b46a00d2259b46753859b36360a5caaeff6a9d1c9e323d84ad4 not found: ID does not exist" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.458040 5035 scope.go:117] "RemoveContainer" containerID="648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b" Nov 24 20:12:55 crc kubenswrapper[5035]: E1124 20:12:55.458526 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b\": container with ID starting with 648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b not found: ID does not exist" containerID="648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.458573 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b"} err="failed to get container status \"648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b\": rpc error: code = NotFound desc = could not find container \"648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b\": container with ID starting with 648458866b20944d9d8133afffce8cf4adfbe6e18bc03cb5c3fe537215f8856b not found: ID does not exist" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.458595 5035 scope.go:117] "RemoveContainer" containerID="c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c" Nov 24 20:12:55 crc kubenswrapper[5035]: E1124 20:12:55.458835 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c\": container with ID starting with c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c not found: ID does not exist" containerID="c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c" Nov 24 20:12:55 crc kubenswrapper[5035]: I1124 20:12:55.458855 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c"} err="failed to get container status \"c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c\": rpc error: code = NotFound desc = could not find container \"c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c\": container with ID starting with c39819bfc22b7a95d904cb42880fd5eb680a12fbd7c17cfa7da8676dbd848f7c not found: ID does not exist" Nov 24 20:12:56 crc kubenswrapper[5035]: I1124 20:12:56.218182 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" path="/var/lib/kubelet/pods/ada10e45-c886-41a0-a6b4-2d16debe62a7/volumes" Nov 24 20:13:06 crc kubenswrapper[5035]: E1124 20:13:06.661661 5035 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.47:43730->38.102.83.47:43781: write tcp 38.102.83.47:43730->38.102.83.47:43781: write: broken pipe Nov 24 20:13:29 crc kubenswrapper[5035]: I1124 20:13:29.922135 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q"] Nov 24 20:13:29 crc kubenswrapper[5035]: E1124 20:13:29.923968 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="extract-utilities" Nov 24 20:13:29 crc kubenswrapper[5035]: I1124 20:13:29.924045 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="extract-utilities" Nov 24 20:13:29 crc kubenswrapper[5035]: E1124 20:13:29.924103 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="registry-server" Nov 24 20:13:29 crc kubenswrapper[5035]: I1124 20:13:29.924155 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="registry-server" Nov 24 20:13:29 crc kubenswrapper[5035]: E1124 20:13:29.924215 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="extract-content" Nov 24 20:13:29 crc kubenswrapper[5035]: I1124 20:13:29.924269 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="extract-content" Nov 24 20:13:29 crc kubenswrapper[5035]: I1124 20:13:29.924527 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ada10e45-c886-41a0-a6b4-2d16debe62a7" containerName="registry-server" Nov 24 20:13:29 crc kubenswrapper[5035]: I1124 20:13:29.925166 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" Nov 24 20:13:29 crc kubenswrapper[5035]: I1124 20:13:29.949862 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q"] Nov 24 20:13:30 crc kubenswrapper[5035]: I1124 20:13:30.083121 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf7nd\" (UniqueName: \"kubernetes.io/projected/b9888ed0-83e5-418d-9187-564b3e4832eb-kube-api-access-lf7nd\") pod \"openstack-operator-controller-operator-556f94c7bb-bdq4q\" (UID: \"b9888ed0-83e5-418d-9187-564b3e4832eb\") " pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" Nov 24 20:13:30 crc kubenswrapper[5035]: I1124 20:13:30.184817 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf7nd\" (UniqueName: \"kubernetes.io/projected/b9888ed0-83e5-418d-9187-564b3e4832eb-kube-api-access-lf7nd\") pod \"openstack-operator-controller-operator-556f94c7bb-bdq4q\" (UID: \"b9888ed0-83e5-418d-9187-564b3e4832eb\") " pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" Nov 24 20:13:30 crc kubenswrapper[5035]: I1124 20:13:30.210832 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf7nd\" (UniqueName: \"kubernetes.io/projected/b9888ed0-83e5-418d-9187-564b3e4832eb-kube-api-access-lf7nd\") pod \"openstack-operator-controller-operator-556f94c7bb-bdq4q\" (UID: \"b9888ed0-83e5-418d-9187-564b3e4832eb\") " pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" Nov 24 20:13:30 crc kubenswrapper[5035]: I1124 20:13:30.248233 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" Nov 24 20:13:30 crc kubenswrapper[5035]: I1124 20:13:30.774997 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q"] Nov 24 20:13:31 crc kubenswrapper[5035]: I1124 20:13:31.750058 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" event={"ID":"b9888ed0-83e5-418d-9187-564b3e4832eb","Type":"ContainerStarted","Data":"8e6eec56699b96fe0a7b03fa1353acc660f16b335838c749175f825d77a43113"} Nov 24 20:13:31 crc kubenswrapper[5035]: I1124 20:13:31.751603 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" event={"ID":"b9888ed0-83e5-418d-9187-564b3e4832eb","Type":"ContainerStarted","Data":"3f24d22681b31fdec855e7283193165c1ea5dfe04d06b81a36aa293ee3f51641"} Nov 24 20:13:31 crc kubenswrapper[5035]: I1124 20:13:31.752882 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" Nov 24 20:13:31 crc kubenswrapper[5035]: I1124 20:13:31.803330 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" podStartSLOduration=2.803281024 podStartE2EDuration="2.803281024s" podCreationTimestamp="2025-11-24 20:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:13:31.794218919 +0000 UTC m=+3550.316725236" watchObservedRunningTime="2025-11-24 20:13:31.803281024 +0000 UTC m=+3550.325787311" Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.251420 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-556f94c7bb-bdq4q" Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.408105 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q"] Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.412473 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" podUID="e33191d3-0eda-4de9-80d6-cd5d077d1862" containerName="operator" containerID="cri-o://3d257b51d15f63f9b8e4558df03bdc2cea33f02fba8ec4c3a9b862d47e01c619" gracePeriod=10 Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.831428 5035 generic.go:334] "Generic (PLEG): container finished" podID="e33191d3-0eda-4de9-80d6-cd5d077d1862" containerID="3d257b51d15f63f9b8e4558df03bdc2cea33f02fba8ec4c3a9b862d47e01c619" exitCode=0 Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.831485 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" event={"ID":"e33191d3-0eda-4de9-80d6-cd5d077d1862","Type":"ContainerDied","Data":"3d257b51d15f63f9b8e4558df03bdc2cea33f02fba8ec4c3a9b862d47e01c619"} Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.831886 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" event={"ID":"e33191d3-0eda-4de9-80d6-cd5d077d1862","Type":"ContainerDied","Data":"3ad7983857682722e58494ce2552fc21c7c4b0231239892f1e36cccf7f552b2c"} Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.831910 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ad7983857682722e58494ce2552fc21c7c4b0231239892f1e36cccf7f552b2c" Nov 24 20:13:40 crc kubenswrapper[5035]: I1124 20:13:40.878761 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 20:13:41 crc kubenswrapper[5035]: I1124 20:13:41.059419 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zmvb\" (UniqueName: \"kubernetes.io/projected/e33191d3-0eda-4de9-80d6-cd5d077d1862-kube-api-access-5zmvb\") pod \"e33191d3-0eda-4de9-80d6-cd5d077d1862\" (UID: \"e33191d3-0eda-4de9-80d6-cd5d077d1862\") " Nov 24 20:13:41 crc kubenswrapper[5035]: I1124 20:13:41.065547 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e33191d3-0eda-4de9-80d6-cd5d077d1862-kube-api-access-5zmvb" (OuterVolumeSpecName: "kube-api-access-5zmvb") pod "e33191d3-0eda-4de9-80d6-cd5d077d1862" (UID: "e33191d3-0eda-4de9-80d6-cd5d077d1862"). InnerVolumeSpecName "kube-api-access-5zmvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:13:41 crc kubenswrapper[5035]: I1124 20:13:41.161890 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zmvb\" (UniqueName: \"kubernetes.io/projected/e33191d3-0eda-4de9-80d6-cd5d077d1862-kube-api-access-5zmvb\") on node \"crc\" DevicePath \"\"" Nov 24 20:13:41 crc kubenswrapper[5035]: I1124 20:13:41.839875 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q" Nov 24 20:13:41 crc kubenswrapper[5035]: I1124 20:13:41.876320 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q"] Nov 24 20:13:41 crc kubenswrapper[5035]: I1124 20:13:41.884368 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-6tq5q"] Nov 24 20:13:42 crc kubenswrapper[5035]: I1124 20:13:42.221037 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e33191d3-0eda-4de9-80d6-cd5d077d1862" path="/var/lib/kubelet/pods/e33191d3-0eda-4de9-80d6-cd5d077d1862/volumes" Nov 24 20:13:45 crc kubenswrapper[5035]: I1124 20:13:45.234049 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:13:45 crc kubenswrapper[5035]: I1124 20:13:45.234624 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:14:15 crc kubenswrapper[5035]: I1124 20:14:15.234752 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:14:15 crc kubenswrapper[5035]: I1124 20:14:15.235748 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.465976 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp"] Nov 24 20:14:16 crc kubenswrapper[5035]: E1124 20:14:16.466773 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33191d3-0eda-4de9-80d6-cd5d077d1862" containerName="operator" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.466812 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33191d3-0eda-4de9-80d6-cd5d077d1862" containerName="operator" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.467108 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e33191d3-0eda-4de9-80d6-cd5d077d1862" containerName="operator" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.468988 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.492271 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp"] Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.544436 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54zjj\" (UniqueName: \"kubernetes.io/projected/b675c0a3-b86b-4993-902e-7a6f041a907a-kube-api-access-54zjj\") pod \"test-operator-controller-manager-69bf7c8699-f9hjp\" (UID: \"b675c0a3-b86b-4993-902e-7a6f041a907a\") " pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.647269 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54zjj\" (UniqueName: \"kubernetes.io/projected/b675c0a3-b86b-4993-902e-7a6f041a907a-kube-api-access-54zjj\") pod \"test-operator-controller-manager-69bf7c8699-f9hjp\" (UID: \"b675c0a3-b86b-4993-902e-7a6f041a907a\") " pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.666794 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54zjj\" (UniqueName: \"kubernetes.io/projected/b675c0a3-b86b-4993-902e-7a6f041a907a-kube-api-access-54zjj\") pod \"test-operator-controller-manager-69bf7c8699-f9hjp\" (UID: \"b675c0a3-b86b-4993-902e-7a6f041a907a\") " pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" Nov 24 20:14:16 crc kubenswrapper[5035]: I1124 20:14:16.812826 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" Nov 24 20:14:17 crc kubenswrapper[5035]: I1124 20:14:17.306450 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp"] Nov 24 20:14:18 crc kubenswrapper[5035]: I1124 20:14:18.264136 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" event={"ID":"b675c0a3-b86b-4993-902e-7a6f041a907a","Type":"ContainerStarted","Data":"042671dff5d81a20832d0dd45bbe769515efc20a951e902582ae13531f080558"} Nov 24 20:14:19 crc kubenswrapper[5035]: I1124 20:14:19.289679 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" event={"ID":"b675c0a3-b86b-4993-902e-7a6f041a907a","Type":"ContainerStarted","Data":"634fe1066ced99978dd7153906cdb2fe02a440a0dabe9ae4696871da9a343e44"} Nov 24 20:14:19 crc kubenswrapper[5035]: I1124 20:14:19.290311 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" Nov 24 20:14:19 crc kubenswrapper[5035]: I1124 20:14:19.290360 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" event={"ID":"b675c0a3-b86b-4993-902e-7a6f041a907a","Type":"ContainerStarted","Data":"ded0a05892c050cc6085d21f0b05d90906814eefcea29dfdf04aaac9c97dc151"} Nov 24 20:14:19 crc kubenswrapper[5035]: I1124 20:14:19.320263 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" podStartSLOduration=2.316476266 podStartE2EDuration="3.320233583s" podCreationTimestamp="2025-11-24 20:14:16 +0000 UTC" firstStartedPulling="2025-11-24 20:14:17.323089067 +0000 UTC m=+3595.845595364" lastFinishedPulling="2025-11-24 20:14:18.326846424 +0000 UTC m=+3596.849352681" observedRunningTime="2025-11-24 20:14:19.312449244 +0000 UTC m=+3597.834955511" watchObservedRunningTime="2025-11-24 20:14:19.320233583 +0000 UTC m=+3597.842739860" Nov 24 20:14:26 crc kubenswrapper[5035]: I1124 20:14:26.816448 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-69bf7c8699-f9hjp" Nov 24 20:14:26 crc kubenswrapper[5035]: I1124 20:14:26.886080 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm"] Nov 24 20:14:26 crc kubenswrapper[5035]: I1124 20:14:26.886355 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="manager" containerID="cri-o://36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a" gracePeriod=10 Nov 24 20:14:26 crc kubenswrapper[5035]: I1124 20:14:26.886513 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="kube-rbac-proxy" containerID="cri-o://b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2" gracePeriod=10 Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.324256 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.380031 5035 generic.go:334] "Generic (PLEG): container finished" podID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerID="b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2" exitCode=0 Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.380073 5035 generic.go:334] "Generic (PLEG): container finished" podID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerID="36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a" exitCode=0 Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.380099 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" event={"ID":"5d0cb726-bb99-44fe-97c5-000e9619a714","Type":"ContainerDied","Data":"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2"} Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.380132 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" event={"ID":"5d0cb726-bb99-44fe-97c5-000e9619a714","Type":"ContainerDied","Data":"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a"} Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.380098 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.380145 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm" event={"ID":"5d0cb726-bb99-44fe-97c5-000e9619a714","Type":"ContainerDied","Data":"e9fb9da303cabc9e96dcd8513b359358eaab0fa7d64e7a8f4673297c0366eab6"} Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.380165 5035 scope.go:117] "RemoveContainer" containerID="b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.399408 5035 scope.go:117] "RemoveContainer" containerID="36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.417159 5035 scope.go:117] "RemoveContainer" containerID="b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2" Nov 24 20:14:27 crc kubenswrapper[5035]: E1124 20:14:27.417674 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2\": container with ID starting with b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2 not found: ID does not exist" containerID="b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.417733 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2"} err="failed to get container status \"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2\": rpc error: code = NotFound desc = could not find container \"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2\": container with ID starting with b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2 not found: ID does not exist" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.417768 5035 scope.go:117] "RemoveContainer" containerID="36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a" Nov 24 20:14:27 crc kubenswrapper[5035]: E1124 20:14:27.418076 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a\": container with ID starting with 36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a not found: ID does not exist" containerID="36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.418109 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a"} err="failed to get container status \"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a\": rpc error: code = NotFound desc = could not find container \"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a\": container with ID starting with 36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a not found: ID does not exist" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.418134 5035 scope.go:117] "RemoveContainer" containerID="b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.418560 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2"} err="failed to get container status \"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2\": rpc error: code = NotFound desc = could not find container \"b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2\": container with ID starting with b2b28014d3c31bb8434b1f2fcec5e054df7642c4e660ed126ad51978b95bc5e2 not found: ID does not exist" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.418598 5035 scope.go:117] "RemoveContainer" containerID="36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.419011 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a"} err="failed to get container status \"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a\": rpc error: code = NotFound desc = could not find container \"36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a\": container with ID starting with 36af01b211f859b3c3fdd1ba55d8f489415bc9d4c8e83595a14661cf777e804a not found: ID does not exist" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.481633 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l24pf\" (UniqueName: \"kubernetes.io/projected/5d0cb726-bb99-44fe-97c5-000e9619a714-kube-api-access-l24pf\") pod \"5d0cb726-bb99-44fe-97c5-000e9619a714\" (UID: \"5d0cb726-bb99-44fe-97c5-000e9619a714\") " Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.488136 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d0cb726-bb99-44fe-97c5-000e9619a714-kube-api-access-l24pf" (OuterVolumeSpecName: "kube-api-access-l24pf") pod "5d0cb726-bb99-44fe-97c5-000e9619a714" (UID: "5d0cb726-bb99-44fe-97c5-000e9619a714"). InnerVolumeSpecName "kube-api-access-l24pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.584553 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l24pf\" (UniqueName: \"kubernetes.io/projected/5d0cb726-bb99-44fe-97c5-000e9619a714-kube-api-access-l24pf\") on node \"crc\" DevicePath \"\"" Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.722365 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm"] Nov 24 20:14:27 crc kubenswrapper[5035]: I1124 20:14:27.729746 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-vjkfm"] Nov 24 20:14:28 crc kubenswrapper[5035]: I1124 20:14:28.220067 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" path="/var/lib/kubelet/pods/5d0cb726-bb99-44fe-97c5-000e9619a714/volumes" Nov 24 20:14:30 crc kubenswrapper[5035]: I1124 20:14:30.689099 5035 scope.go:117] "RemoveContainer" containerID="3d257b51d15f63f9b8e4558df03bdc2cea33f02fba8ec4c3a9b862d47e01c619" Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.234453 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.235021 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.235075 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.236061 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.236126 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" gracePeriod=600 Nov 24 20:14:45 crc kubenswrapper[5035]: E1124 20:14:45.359784 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.609812 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" exitCode=0 Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.609858 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1"} Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.609899 5035 scope.go:117] "RemoveContainer" containerID="cbcdbc5dae82463d722971b6c0e1597afa8425f378dbbe8431940fc4c08a86af" Nov 24 20:14:45 crc kubenswrapper[5035]: I1124 20:14:45.610815 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:14:45 crc kubenswrapper[5035]: E1124 20:14:45.611337 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:14:58 crc kubenswrapper[5035]: I1124 20:14:58.202594 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:14:58 crc kubenswrapper[5035]: E1124 20:14:58.204284 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.192778 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj"] Nov 24 20:15:00 crc kubenswrapper[5035]: E1124 20:15:00.193911 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="manager" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.193938 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="manager" Nov 24 20:15:00 crc kubenswrapper[5035]: E1124 20:15:00.194005 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="kube-rbac-proxy" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.194018 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="kube-rbac-proxy" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.194395 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="kube-rbac-proxy" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.194489 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d0cb726-bb99-44fe-97c5-000e9619a714" containerName="manager" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.195730 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.197925 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.199346 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.230082 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj"] Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.259431 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ae512c2-b562-4020-822f-c3e97c4caf95-secret-volume\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.259924 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsqlf\" (UniqueName: \"kubernetes.io/projected/8ae512c2-b562-4020-822f-c3e97c4caf95-kube-api-access-lsqlf\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.260076 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ae512c2-b562-4020-822f-c3e97c4caf95-config-volume\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.363220 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ae512c2-b562-4020-822f-c3e97c4caf95-secret-volume\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.363492 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsqlf\" (UniqueName: \"kubernetes.io/projected/8ae512c2-b562-4020-822f-c3e97c4caf95-kube-api-access-lsqlf\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.363547 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ae512c2-b562-4020-822f-c3e97c4caf95-config-volume\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.365114 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ae512c2-b562-4020-822f-c3e97c4caf95-config-volume\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.372921 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ae512c2-b562-4020-822f-c3e97c4caf95-secret-volume\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.395159 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsqlf\" (UniqueName: \"kubernetes.io/projected/8ae512c2-b562-4020-822f-c3e97c4caf95-kube-api-access-lsqlf\") pod \"collect-profiles-29400255-vwsmj\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:00 crc kubenswrapper[5035]: I1124 20:15:00.537742 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:01 crc kubenswrapper[5035]: I1124 20:15:01.035053 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj"] Nov 24 20:15:01 crc kubenswrapper[5035]: I1124 20:15:01.805004 5035 generic.go:334] "Generic (PLEG): container finished" podID="8ae512c2-b562-4020-822f-c3e97c4caf95" containerID="371bb4464f503221e5dff224c0edd7a25a8e8534facc83920940756c55dba90c" exitCode=0 Nov 24 20:15:01 crc kubenswrapper[5035]: I1124 20:15:01.805085 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" event={"ID":"8ae512c2-b562-4020-822f-c3e97c4caf95","Type":"ContainerDied","Data":"371bb4464f503221e5dff224c0edd7a25a8e8534facc83920940756c55dba90c"} Nov 24 20:15:01 crc kubenswrapper[5035]: I1124 20:15:01.805569 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" event={"ID":"8ae512c2-b562-4020-822f-c3e97c4caf95","Type":"ContainerStarted","Data":"18570dd9ef46a4bb53d605393c2a7aa2c82f713e02cf718d3c079127f6b41057"} Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.217216 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.335662 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsqlf\" (UniqueName: \"kubernetes.io/projected/8ae512c2-b562-4020-822f-c3e97c4caf95-kube-api-access-lsqlf\") pod \"8ae512c2-b562-4020-822f-c3e97c4caf95\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.335784 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ae512c2-b562-4020-822f-c3e97c4caf95-config-volume\") pod \"8ae512c2-b562-4020-822f-c3e97c4caf95\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.335816 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ae512c2-b562-4020-822f-c3e97c4caf95-secret-volume\") pod \"8ae512c2-b562-4020-822f-c3e97c4caf95\" (UID: \"8ae512c2-b562-4020-822f-c3e97c4caf95\") " Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.337108 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ae512c2-b562-4020-822f-c3e97c4caf95-config-volume" (OuterVolumeSpecName: "config-volume") pod "8ae512c2-b562-4020-822f-c3e97c4caf95" (UID: "8ae512c2-b562-4020-822f-c3e97c4caf95"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.345037 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae512c2-b562-4020-822f-c3e97c4caf95-kube-api-access-lsqlf" (OuterVolumeSpecName: "kube-api-access-lsqlf") pod "8ae512c2-b562-4020-822f-c3e97c4caf95" (UID: "8ae512c2-b562-4020-822f-c3e97c4caf95"). InnerVolumeSpecName "kube-api-access-lsqlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.355464 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ae512c2-b562-4020-822f-c3e97c4caf95-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8ae512c2-b562-4020-822f-c3e97c4caf95" (UID: "8ae512c2-b562-4020-822f-c3e97c4caf95"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.438184 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ae512c2-b562-4020-822f-c3e97c4caf95-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.438541 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ae512c2-b562-4020-822f-c3e97c4caf95-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.438557 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsqlf\" (UniqueName: \"kubernetes.io/projected/8ae512c2-b562-4020-822f-c3e97c4caf95-kube-api-access-lsqlf\") on node \"crc\" DevicePath \"\"" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.827461 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" event={"ID":"8ae512c2-b562-4020-822f-c3e97c4caf95","Type":"ContainerDied","Data":"18570dd9ef46a4bb53d605393c2a7aa2c82f713e02cf718d3c079127f6b41057"} Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.827521 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18570dd9ef46a4bb53d605393c2a7aa2c82f713e02cf718d3c079127f6b41057" Nov 24 20:15:03 crc kubenswrapper[5035]: I1124 20:15:03.827549 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj" Nov 24 20:15:04 crc kubenswrapper[5035]: I1124 20:15:04.300332 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w"] Nov 24 20:15:04 crc kubenswrapper[5035]: I1124 20:15:04.309749 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400210-26p8w"] Nov 24 20:15:06 crc kubenswrapper[5035]: I1124 20:15:06.220211 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e46e48-d877-4332-918a-95d8c6afc3b5" path="/var/lib/kubelet/pods/b3e46e48-d877-4332-918a-95d8c6afc3b5/volumes" Nov 24 20:15:13 crc kubenswrapper[5035]: I1124 20:15:13.200709 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:15:13 crc kubenswrapper[5035]: E1124 20:15:13.201477 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:15:28 crc kubenswrapper[5035]: I1124 20:15:28.200967 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:15:28 crc kubenswrapper[5035]: E1124 20:15:28.202167 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:15:30 crc kubenswrapper[5035]: I1124 20:15:30.815418 5035 scope.go:117] "RemoveContainer" containerID="7a4970c6567426499877620ccac9fa6bcb84b68ebdfaaab1c0abc47abb90c7dc" Nov 24 20:15:40 crc kubenswrapper[5035]: I1124 20:15:40.202064 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:15:40 crc kubenswrapper[5035]: E1124 20:15:40.204818 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:15:55 crc kubenswrapper[5035]: I1124 20:15:55.200392 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:15:55 crc kubenswrapper[5035]: E1124 20:15:55.201514 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:16:06 crc kubenswrapper[5035]: I1124 20:16:06.200910 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:16:06 crc kubenswrapper[5035]: E1124 20:16:06.202121 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.200324 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:16:21 crc kubenswrapper[5035]: E1124 20:16:21.201179 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.296561 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Nov 24 20:16:21 crc kubenswrapper[5035]: E1124 20:16:21.297281 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae512c2-b562-4020-822f-c3e97c4caf95" containerName="collect-profiles" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.297325 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae512c2-b562-4020-822f-c3e97c4caf95" containerName="collect-profiles" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.297537 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae512c2-b562-4020-822f-c3e97c4caf95" containerName="collect-profiles" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.298317 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.305118 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.305361 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.305444 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.305589 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bz72s" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.319622 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463033 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463113 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463164 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463394 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463441 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463476 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6rxt\" (UniqueName: \"kubernetes.io/projected/03d6dde6-e977-4ec1-a23d-1642e6c23029-kube-api-access-j6rxt\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463693 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.463777 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.464058 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.464187 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566277 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566443 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566513 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566600 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566612 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566726 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566780 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566854 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.566952 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6rxt\" (UniqueName: \"kubernetes.io/projected/03d6dde6-e977-4ec1-a23d-1642e6c23029-kube-api-access-j6rxt\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.567073 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.567154 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.567927 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.569100 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.569742 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.578775 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.582025 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.582941 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.583870 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.583951 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.632214 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6rxt\" (UniqueName: \"kubernetes.io/projected/03d6dde6-e977-4ec1-a23d-1642e6c23029-kube-api-access-j6rxt\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.633285 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:21 crc kubenswrapper[5035]: I1124 20:16:21.934040 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:16:22 crc kubenswrapper[5035]: I1124 20:16:22.532203 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Nov 24 20:16:22 crc kubenswrapper[5035]: I1124 20:16:22.726488 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"03d6dde6-e977-4ec1-a23d-1642e6c23029","Type":"ContainerStarted","Data":"25d4449e6036685fad57b15761165d9327cea44573d1ab2ce44a350f68963d09"} Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.729799 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9m4dr"] Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.733954 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.740406 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9m4dr"] Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.828860 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-catalog-content\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.829182 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtbkh\" (UniqueName: \"kubernetes.io/projected/74cc1e96-0942-4845-ba11-3106cd183f9c-kube-api-access-xtbkh\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.829260 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-utilities\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.931059 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtbkh\" (UniqueName: \"kubernetes.io/projected/74cc1e96-0942-4845-ba11-3106cd183f9c-kube-api-access-xtbkh\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.931116 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-utilities\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.931189 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-catalog-content\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.931702 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-catalog-content\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.932191 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-utilities\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:27 crc kubenswrapper[5035]: I1124 20:16:27.953077 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtbkh\" (UniqueName: \"kubernetes.io/projected/74cc1e96-0942-4845-ba11-3106cd183f9c-kube-api-access-xtbkh\") pod \"certified-operators-9m4dr\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:28 crc kubenswrapper[5035]: I1124 20:16:28.056095 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:28 crc kubenswrapper[5035]: I1124 20:16:28.655936 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9m4dr"] Nov 24 20:16:28 crc kubenswrapper[5035]: I1124 20:16:28.784255 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4dr" event={"ID":"74cc1e96-0942-4845-ba11-3106cd183f9c","Type":"ContainerStarted","Data":"3fc3632ff05da7d2a6b1320d24c36cf4297361b0e5072331c4dd06df48a755b0"} Nov 24 20:16:29 crc kubenswrapper[5035]: I1124 20:16:29.799878 5035 generic.go:334] "Generic (PLEG): container finished" podID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerID="1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf" exitCode=0 Nov 24 20:16:29 crc kubenswrapper[5035]: I1124 20:16:29.799961 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4dr" event={"ID":"74cc1e96-0942-4845-ba11-3106cd183f9c","Type":"ContainerDied","Data":"1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf"} Nov 24 20:16:31 crc kubenswrapper[5035]: I1124 20:16:31.823839 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4dr" event={"ID":"74cc1e96-0942-4845-ba11-3106cd183f9c","Type":"ContainerStarted","Data":"2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc"} Nov 24 20:16:33 crc kubenswrapper[5035]: I1124 20:16:33.856897 5035 generic.go:334] "Generic (PLEG): container finished" podID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerID="2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc" exitCode=0 Nov 24 20:16:33 crc kubenswrapper[5035]: I1124 20:16:33.856987 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4dr" event={"ID":"74cc1e96-0942-4845-ba11-3106cd183f9c","Type":"ContainerDied","Data":"2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc"} Nov 24 20:16:34 crc kubenswrapper[5035]: I1124 20:16:34.200723 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:16:34 crc kubenswrapper[5035]: E1124 20:16:34.201677 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:16:47 crc kubenswrapper[5035]: I1124 20:16:47.200912 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:16:47 crc kubenswrapper[5035]: E1124 20:16:47.201814 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:16:54 crc kubenswrapper[5035]: E1124 20:16:54.372706 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 24 20:16:54 crc kubenswrapper[5035]: E1124 20:16:54.373903 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j6rxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-full_openstack(03d6dde6-e977-4ec1-a23d-1642e6c23029): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 20:16:54 crc kubenswrapper[5035]: E1124 20:16:54.375156 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="03d6dde6-e977-4ec1-a23d-1642e6c23029" Nov 24 20:16:55 crc kubenswrapper[5035]: E1124 20:16:55.217369 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="03d6dde6-e977-4ec1-a23d-1642e6c23029" Nov 24 20:16:56 crc kubenswrapper[5035]: I1124 20:16:56.224155 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4dr" event={"ID":"74cc1e96-0942-4845-ba11-3106cd183f9c","Type":"ContainerStarted","Data":"ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6"} Nov 24 20:16:56 crc kubenswrapper[5035]: I1124 20:16:56.256467 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9m4dr" podStartSLOduration=4.388018766 podStartE2EDuration="29.256441537s" podCreationTimestamp="2025-11-24 20:16:27 +0000 UTC" firstStartedPulling="2025-11-24 20:16:29.802376752 +0000 UTC m=+3728.324883049" lastFinishedPulling="2025-11-24 20:16:54.670799553 +0000 UTC m=+3753.193305820" observedRunningTime="2025-11-24 20:16:56.247570089 +0000 UTC m=+3754.770076356" watchObservedRunningTime="2025-11-24 20:16:56.256441537 +0000 UTC m=+3754.778947794" Nov 24 20:16:58 crc kubenswrapper[5035]: I1124 20:16:58.057011 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:58 crc kubenswrapper[5035]: I1124 20:16:58.057393 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:16:59 crc kubenswrapper[5035]: I1124 20:16:59.128699 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9m4dr" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="registry-server" probeResult="failure" output=< Nov 24 20:16:59 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:16:59 crc kubenswrapper[5035]: > Nov 24 20:17:00 crc kubenswrapper[5035]: I1124 20:17:00.200032 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:17:00 crc kubenswrapper[5035]: E1124 20:17:00.200582 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:17:06 crc kubenswrapper[5035]: I1124 20:17:06.869897 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 20:17:08 crc kubenswrapper[5035]: I1124 20:17:08.354249 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"03d6dde6-e977-4ec1-a23d-1642e6c23029","Type":"ContainerStarted","Data":"06c50d2f475343fb268f78f8b509ef1a07f252e6f5c5d715bf3d531053913a30"} Nov 24 20:17:08 crc kubenswrapper[5035]: I1124 20:17:08.385675 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-full" podStartSLOduration=4.057430503 podStartE2EDuration="48.385655431s" podCreationTimestamp="2025-11-24 20:16:20 +0000 UTC" firstStartedPulling="2025-11-24 20:16:22.539113027 +0000 UTC m=+3721.061619284" lastFinishedPulling="2025-11-24 20:17:06.867337935 +0000 UTC m=+3765.389844212" observedRunningTime="2025-11-24 20:17:08.376566567 +0000 UTC m=+3766.899072834" watchObservedRunningTime="2025-11-24 20:17:08.385655431 +0000 UTC m=+3766.908161688" Nov 24 20:17:09 crc kubenswrapper[5035]: I1124 20:17:09.114038 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9m4dr" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="registry-server" probeResult="failure" output=< Nov 24 20:17:09 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:17:09 crc kubenswrapper[5035]: > Nov 24 20:17:14 crc kubenswrapper[5035]: I1124 20:17:14.200782 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:17:14 crc kubenswrapper[5035]: E1124 20:17:14.202446 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:17:18 crc kubenswrapper[5035]: I1124 20:17:18.114785 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:17:18 crc kubenswrapper[5035]: I1124 20:17:18.179654 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:17:18 crc kubenswrapper[5035]: I1124 20:17:18.360736 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9m4dr"] Nov 24 20:17:19 crc kubenswrapper[5035]: I1124 20:17:19.485032 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9m4dr" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="registry-server" containerID="cri-o://ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6" gracePeriod=2 Nov 24 20:17:19 crc kubenswrapper[5035]: I1124 20:17:19.973874 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.116593 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtbkh\" (UniqueName: \"kubernetes.io/projected/74cc1e96-0942-4845-ba11-3106cd183f9c-kube-api-access-xtbkh\") pod \"74cc1e96-0942-4845-ba11-3106cd183f9c\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.116689 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-catalog-content\") pod \"74cc1e96-0942-4845-ba11-3106cd183f9c\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.116813 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-utilities\") pod \"74cc1e96-0942-4845-ba11-3106cd183f9c\" (UID: \"74cc1e96-0942-4845-ba11-3106cd183f9c\") " Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.117856 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-utilities" (OuterVolumeSpecName: "utilities") pod "74cc1e96-0942-4845-ba11-3106cd183f9c" (UID: "74cc1e96-0942-4845-ba11-3106cd183f9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.127605 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74cc1e96-0942-4845-ba11-3106cd183f9c-kube-api-access-xtbkh" (OuterVolumeSpecName: "kube-api-access-xtbkh") pod "74cc1e96-0942-4845-ba11-3106cd183f9c" (UID: "74cc1e96-0942-4845-ba11-3106cd183f9c"). InnerVolumeSpecName "kube-api-access-xtbkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.170676 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74cc1e96-0942-4845-ba11-3106cd183f9c" (UID: "74cc1e96-0942-4845-ba11-3106cd183f9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.219956 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.220010 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74cc1e96-0942-4845-ba11-3106cd183f9c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.220032 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtbkh\" (UniqueName: \"kubernetes.io/projected/74cc1e96-0942-4845-ba11-3106cd183f9c-kube-api-access-xtbkh\") on node \"crc\" DevicePath \"\"" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.503049 5035 generic.go:334] "Generic (PLEG): container finished" podID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerID="ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6" exitCode=0 Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.503118 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4dr" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.503129 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4dr" event={"ID":"74cc1e96-0942-4845-ba11-3106cd183f9c","Type":"ContainerDied","Data":"ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6"} Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.503183 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4dr" event={"ID":"74cc1e96-0942-4845-ba11-3106cd183f9c","Type":"ContainerDied","Data":"3fc3632ff05da7d2a6b1320d24c36cf4297361b0e5072331c4dd06df48a755b0"} Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.503202 5035 scope.go:117] "RemoveContainer" containerID="ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.525878 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9m4dr"] Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.537232 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9m4dr"] Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.557015 5035 scope.go:117] "RemoveContainer" containerID="2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.585765 5035 scope.go:117] "RemoveContainer" containerID="1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.656900 5035 scope.go:117] "RemoveContainer" containerID="ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6" Nov 24 20:17:20 crc kubenswrapper[5035]: E1124 20:17:20.667820 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6\": container with ID starting with ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6 not found: ID does not exist" containerID="ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.667865 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6"} err="failed to get container status \"ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6\": rpc error: code = NotFound desc = could not find container \"ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6\": container with ID starting with ef4e4171f7e41e921e186ad5a0acd12e4d203de5d06a66d1cb3fd220ba5d71f6 not found: ID does not exist" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.667891 5035 scope.go:117] "RemoveContainer" containerID="2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc" Nov 24 20:17:20 crc kubenswrapper[5035]: E1124 20:17:20.669138 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc\": container with ID starting with 2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc not found: ID does not exist" containerID="2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.669176 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc"} err="failed to get container status \"2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc\": rpc error: code = NotFound desc = could not find container \"2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc\": container with ID starting with 2e52a24bcccad1ee47c0256545a7a31438b48fe469900c41ab681f70be2bfafc not found: ID does not exist" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.669200 5035 scope.go:117] "RemoveContainer" containerID="1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf" Nov 24 20:17:20 crc kubenswrapper[5035]: E1124 20:17:20.669874 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf\": container with ID starting with 1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf not found: ID does not exist" containerID="1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf" Nov 24 20:17:20 crc kubenswrapper[5035]: I1124 20:17:20.669907 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf"} err="failed to get container status \"1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf\": rpc error: code = NotFound desc = could not find container \"1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf\": container with ID starting with 1ee82d6f4d3775dcd285d9a60a972b4b872d4c74335774aec9534e4a453611bf not found: ID does not exist" Nov 24 20:17:22 crc kubenswrapper[5035]: I1124 20:17:22.217666 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" path="/var/lib/kubelet/pods/74cc1e96-0942-4845-ba11-3106cd183f9c/volumes" Nov 24 20:17:28 crc kubenswrapper[5035]: I1124 20:17:28.200500 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:17:28 crc kubenswrapper[5035]: E1124 20:17:28.201411 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:17:39 crc kubenswrapper[5035]: I1124 20:17:39.200783 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:17:39 crc kubenswrapper[5035]: E1124 20:17:39.201875 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:17:51 crc kubenswrapper[5035]: I1124 20:17:51.200761 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:17:51 crc kubenswrapper[5035]: E1124 20:17:51.201532 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:18:03 crc kubenswrapper[5035]: I1124 20:18:03.200954 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:18:03 crc kubenswrapper[5035]: E1124 20:18:03.202205 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:18:14 crc kubenswrapper[5035]: I1124 20:18:14.201058 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:18:14 crc kubenswrapper[5035]: E1124 20:18:14.202272 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:18:26 crc kubenswrapper[5035]: I1124 20:18:26.200576 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:18:26 crc kubenswrapper[5035]: E1124 20:18:26.201404 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:18:31 crc kubenswrapper[5035]: I1124 20:18:31.015745 5035 scope.go:117] "RemoveContainer" containerID="b3c154c484a02c79c31a9b661f94344a1bac6d86d124cf810998708d53664246" Nov 24 20:18:31 crc kubenswrapper[5035]: I1124 20:18:31.062049 5035 scope.go:117] "RemoveContainer" containerID="5adebacf6c11e242878dac2f91e8630f27a4a20b724f450e4dac240cca853517" Nov 24 20:18:40 crc kubenswrapper[5035]: I1124 20:18:40.201249 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:18:40 crc kubenswrapper[5035]: E1124 20:18:40.202534 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:18:52 crc kubenswrapper[5035]: I1124 20:18:52.213990 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:18:52 crc kubenswrapper[5035]: E1124 20:18:52.215919 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:19:06 crc kubenswrapper[5035]: I1124 20:19:06.200274 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:19:06 crc kubenswrapper[5035]: E1124 20:19:06.201202 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:19:20 crc kubenswrapper[5035]: I1124 20:19:20.211539 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:19:20 crc kubenswrapper[5035]: E1124 20:19:20.213494 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:19:35 crc kubenswrapper[5035]: I1124 20:19:35.200636 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:19:35 crc kubenswrapper[5035]: E1124 20:19:35.201782 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:19:47 crc kubenswrapper[5035]: I1124 20:19:47.199913 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:19:48 crc kubenswrapper[5035]: I1124 20:19:48.259730 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"ccddfb6bd9389e4f81715231c8b69c39a4065bea00787ec4778a540a35d107d3"} Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.516752 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4fbft"] Nov 24 20:20:21 crc kubenswrapper[5035]: E1124 20:20:21.518356 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="registry-server" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.518375 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="registry-server" Nov 24 20:20:21 crc kubenswrapper[5035]: E1124 20:20:21.518397 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="extract-utilities" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.518404 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="extract-utilities" Nov 24 20:20:21 crc kubenswrapper[5035]: E1124 20:20:21.518445 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="extract-content" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.518453 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="extract-content" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.518679 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="74cc1e96-0942-4845-ba11-3106cd183f9c" containerName="registry-server" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.520495 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.540234 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4fbft"] Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.642768 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-utilities\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.642943 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r4c9\" (UniqueName: \"kubernetes.io/projected/5729c5ed-8650-41ed-85a6-67d1f9804039-kube-api-access-4r4c9\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.643122 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-catalog-content\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.744830 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r4c9\" (UniqueName: \"kubernetes.io/projected/5729c5ed-8650-41ed-85a6-67d1f9804039-kube-api-access-4r4c9\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.744909 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-catalog-content\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.744959 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-utilities\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.745390 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-utilities\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.745564 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-catalog-content\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.769840 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r4c9\" (UniqueName: \"kubernetes.io/projected/5729c5ed-8650-41ed-85a6-67d1f9804039-kube-api-access-4r4c9\") pod \"redhat-operators-4fbft\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:21 crc kubenswrapper[5035]: I1124 20:20:21.859541 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:22 crc kubenswrapper[5035]: I1124 20:20:22.339374 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4fbft"] Nov 24 20:20:22 crc kubenswrapper[5035]: W1124 20:20:22.351863 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5729c5ed_8650_41ed_85a6_67d1f9804039.slice/crio-225ef1e427cdd05b8e8c27c0b23e5ea0d2d46cd26b80b9bfea2aa4d781661697 WatchSource:0}: Error finding container 225ef1e427cdd05b8e8c27c0b23e5ea0d2d46cd26b80b9bfea2aa4d781661697: Status 404 returned error can't find the container with id 225ef1e427cdd05b8e8c27c0b23e5ea0d2d46cd26b80b9bfea2aa4d781661697 Nov 24 20:20:22 crc kubenswrapper[5035]: I1124 20:20:22.641699 5035 generic.go:334] "Generic (PLEG): container finished" podID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerID="d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736" exitCode=0 Nov 24 20:20:22 crc kubenswrapper[5035]: I1124 20:20:22.641741 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fbft" event={"ID":"5729c5ed-8650-41ed-85a6-67d1f9804039","Type":"ContainerDied","Data":"d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736"} Nov 24 20:20:22 crc kubenswrapper[5035]: I1124 20:20:22.641777 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fbft" event={"ID":"5729c5ed-8650-41ed-85a6-67d1f9804039","Type":"ContainerStarted","Data":"225ef1e427cdd05b8e8c27c0b23e5ea0d2d46cd26b80b9bfea2aa4d781661697"} Nov 24 20:20:22 crc kubenswrapper[5035]: I1124 20:20:22.643754 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:20:23 crc kubenswrapper[5035]: I1124 20:20:23.653228 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fbft" event={"ID":"5729c5ed-8650-41ed-85a6-67d1f9804039","Type":"ContainerStarted","Data":"6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa"} Nov 24 20:20:26 crc kubenswrapper[5035]: I1124 20:20:26.688212 5035 generic.go:334] "Generic (PLEG): container finished" podID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerID="6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa" exitCode=0 Nov 24 20:20:26 crc kubenswrapper[5035]: I1124 20:20:26.688278 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fbft" event={"ID":"5729c5ed-8650-41ed-85a6-67d1f9804039","Type":"ContainerDied","Data":"6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa"} Nov 24 20:20:27 crc kubenswrapper[5035]: I1124 20:20:27.702945 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fbft" event={"ID":"5729c5ed-8650-41ed-85a6-67d1f9804039","Type":"ContainerStarted","Data":"345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2"} Nov 24 20:20:27 crc kubenswrapper[5035]: I1124 20:20:27.731939 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4fbft" podStartSLOduration=2.260797296 podStartE2EDuration="6.731911903s" podCreationTimestamp="2025-11-24 20:20:21 +0000 UTC" firstStartedPulling="2025-11-24 20:20:22.643540109 +0000 UTC m=+3961.166046366" lastFinishedPulling="2025-11-24 20:20:27.114654686 +0000 UTC m=+3965.637160973" observedRunningTime="2025-11-24 20:20:27.726782305 +0000 UTC m=+3966.249288622" watchObservedRunningTime="2025-11-24 20:20:27.731911903 +0000 UTC m=+3966.254418190" Nov 24 20:20:31 crc kubenswrapper[5035]: I1124 20:20:31.859745 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:31 crc kubenswrapper[5035]: I1124 20:20:31.860347 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:32 crc kubenswrapper[5035]: I1124 20:20:32.919771 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4fbft" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="registry-server" probeResult="failure" output=< Nov 24 20:20:32 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:20:32 crc kubenswrapper[5035]: > Nov 24 20:20:41 crc kubenswrapper[5035]: I1124 20:20:41.945844 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:42 crc kubenswrapper[5035]: I1124 20:20:42.036363 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:42 crc kubenswrapper[5035]: I1124 20:20:42.214997 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4fbft"] Nov 24 20:20:43 crc kubenswrapper[5035]: I1124 20:20:43.889419 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4fbft" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="registry-server" containerID="cri-o://345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2" gracePeriod=2 Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.398411 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.497671 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r4c9\" (UniqueName: \"kubernetes.io/projected/5729c5ed-8650-41ed-85a6-67d1f9804039-kube-api-access-4r4c9\") pod \"5729c5ed-8650-41ed-85a6-67d1f9804039\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.497974 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-catalog-content\") pod \"5729c5ed-8650-41ed-85a6-67d1f9804039\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.498086 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-utilities\") pod \"5729c5ed-8650-41ed-85a6-67d1f9804039\" (UID: \"5729c5ed-8650-41ed-85a6-67d1f9804039\") " Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.499796 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-utilities" (OuterVolumeSpecName: "utilities") pod "5729c5ed-8650-41ed-85a6-67d1f9804039" (UID: "5729c5ed-8650-41ed-85a6-67d1f9804039"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.504845 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5729c5ed-8650-41ed-85a6-67d1f9804039-kube-api-access-4r4c9" (OuterVolumeSpecName: "kube-api-access-4r4c9") pod "5729c5ed-8650-41ed-85a6-67d1f9804039" (UID: "5729c5ed-8650-41ed-85a6-67d1f9804039"). InnerVolumeSpecName "kube-api-access-4r4c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.592730 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5729c5ed-8650-41ed-85a6-67d1f9804039" (UID: "5729c5ed-8650-41ed-85a6-67d1f9804039"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.601311 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r4c9\" (UniqueName: \"kubernetes.io/projected/5729c5ed-8650-41ed-85a6-67d1f9804039-kube-api-access-4r4c9\") on node \"crc\" DevicePath \"\"" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.601365 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.601385 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5729c5ed-8650-41ed-85a6-67d1f9804039-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.905022 5035 generic.go:334] "Generic (PLEG): container finished" podID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerID="345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2" exitCode=0 Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.905082 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fbft" event={"ID":"5729c5ed-8650-41ed-85a6-67d1f9804039","Type":"ContainerDied","Data":"345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2"} Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.905106 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fbft" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.905153 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fbft" event={"ID":"5729c5ed-8650-41ed-85a6-67d1f9804039","Type":"ContainerDied","Data":"225ef1e427cdd05b8e8c27c0b23e5ea0d2d46cd26b80b9bfea2aa4d781661697"} Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.905189 5035 scope.go:117] "RemoveContainer" containerID="345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.936676 5035 scope.go:117] "RemoveContainer" containerID="6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa" Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.965167 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4fbft"] Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.979039 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4fbft"] Nov 24 20:20:44 crc kubenswrapper[5035]: I1124 20:20:44.995372 5035 scope.go:117] "RemoveContainer" containerID="d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736" Nov 24 20:20:45 crc kubenswrapper[5035]: I1124 20:20:45.060912 5035 scope.go:117] "RemoveContainer" containerID="345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2" Nov 24 20:20:45 crc kubenswrapper[5035]: E1124 20:20:45.061517 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2\": container with ID starting with 345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2 not found: ID does not exist" containerID="345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2" Nov 24 20:20:45 crc kubenswrapper[5035]: I1124 20:20:45.061549 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2"} err="failed to get container status \"345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2\": rpc error: code = NotFound desc = could not find container \"345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2\": container with ID starting with 345c35bbac8e08ecac4ee70471ecc002ef9f4ff26679230dceb2ccd24e42d1e2 not found: ID does not exist" Nov 24 20:20:45 crc kubenswrapper[5035]: I1124 20:20:45.061573 5035 scope.go:117] "RemoveContainer" containerID="6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa" Nov 24 20:20:45 crc kubenswrapper[5035]: E1124 20:20:45.062020 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa\": container with ID starting with 6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa not found: ID does not exist" containerID="6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa" Nov 24 20:20:45 crc kubenswrapper[5035]: I1124 20:20:45.062040 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa"} err="failed to get container status \"6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa\": rpc error: code = NotFound desc = could not find container \"6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa\": container with ID starting with 6c4bd66eb10ab311556ee0c472349aef15dadb25b1a1e05a70e7575dca4412aa not found: ID does not exist" Nov 24 20:20:45 crc kubenswrapper[5035]: I1124 20:20:45.062052 5035 scope.go:117] "RemoveContainer" containerID="d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736" Nov 24 20:20:45 crc kubenswrapper[5035]: E1124 20:20:45.062446 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736\": container with ID starting with d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736 not found: ID does not exist" containerID="d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736" Nov 24 20:20:45 crc kubenswrapper[5035]: I1124 20:20:45.062507 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736"} err="failed to get container status \"d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736\": rpc error: code = NotFound desc = could not find container \"d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736\": container with ID starting with d25c0c6f836e8e881cce90f7c82f6521f2051ada703a626ebab3296c13edf736 not found: ID does not exist" Nov 24 20:20:46 crc kubenswrapper[5035]: I1124 20:20:46.215201 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" path="/var/lib/kubelet/pods/5729c5ed-8650-41ed-85a6-67d1f9804039/volumes" Nov 24 20:21:24 crc kubenswrapper[5035]: I1124 20:21:24.060842 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-65b2-account-create-xpmrl"] Nov 24 20:21:24 crc kubenswrapper[5035]: I1124 20:21:24.072069 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-vk6mj"] Nov 24 20:21:24 crc kubenswrapper[5035]: I1124 20:21:24.079217 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-65b2-account-create-xpmrl"] Nov 24 20:21:24 crc kubenswrapper[5035]: I1124 20:21:24.086184 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-vk6mj"] Nov 24 20:21:24 crc kubenswrapper[5035]: I1124 20:21:24.218761 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7097c70e-55c5-4d34-88a6-fce197f78ba8" path="/var/lib/kubelet/pods/7097c70e-55c5-4d34-88a6-fce197f78ba8/volumes" Nov 24 20:21:24 crc kubenswrapper[5035]: I1124 20:21:24.220140 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7728bae7-82b0-4cf4-824f-4911ff936ed1" path="/var/lib/kubelet/pods/7728bae7-82b0-4cf4-824f-4911ff936ed1/volumes" Nov 24 20:21:31 crc kubenswrapper[5035]: I1124 20:21:31.169285 5035 scope.go:117] "RemoveContainer" containerID="7aba46386f064f79f23c934a770724b2b6674d54aeba739c9a8f18451a3ba143" Nov 24 20:21:31 crc kubenswrapper[5035]: I1124 20:21:31.192821 5035 scope.go:117] "RemoveContainer" containerID="ce26f55c82c86ec494bf4df0fec0dd84d40355744726fcaa595220158601bbdd" Nov 24 20:21:44 crc kubenswrapper[5035]: I1124 20:21:44.032427 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-85wlr"] Nov 24 20:21:44 crc kubenswrapper[5035]: I1124 20:21:44.043621 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-85wlr"] Nov 24 20:21:44 crc kubenswrapper[5035]: I1124 20:21:44.213933 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25" path="/var/lib/kubelet/pods/8b4e36cb-09f2-4f1c-928c-cb0f0a1b3a25/volumes" Nov 24 20:22:15 crc kubenswrapper[5035]: I1124 20:22:15.233949 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:22:15 crc kubenswrapper[5035]: I1124 20:22:15.234562 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:22:31 crc kubenswrapper[5035]: I1124 20:22:31.326911 5035 scope.go:117] "RemoveContainer" containerID="c7642684733f2859e2722ec4b01e9692f1069043d5c03825dfc42e8789eb6538" Nov 24 20:22:45 crc kubenswrapper[5035]: I1124 20:22:45.234059 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:22:45 crc kubenswrapper[5035]: I1124 20:22:45.234719 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.234578 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.235151 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.235214 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.236185 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ccddfb6bd9389e4f81715231c8b69c39a4065bea00787ec4778a540a35d107d3"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.236284 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://ccddfb6bd9389e4f81715231c8b69c39a4065bea00787ec4778a540a35d107d3" gracePeriod=600 Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.455199 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="ccddfb6bd9389e4f81715231c8b69c39a4065bea00787ec4778a540a35d107d3" exitCode=0 Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.455263 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"ccddfb6bd9389e4f81715231c8b69c39a4065bea00787ec4778a540a35d107d3"} Nov 24 20:23:15 crc kubenswrapper[5035]: I1124 20:23:15.455640 5035 scope.go:117] "RemoveContainer" containerID="af38ef40f131036bcd05d5a2b03b04617186d8c8a1d3156387e74537be137ef1" Nov 24 20:23:16 crc kubenswrapper[5035]: I1124 20:23:16.465862 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487"} Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.836820 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6f4sh"] Nov 24 20:23:36 crc kubenswrapper[5035]: E1124 20:23:36.837884 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="registry-server" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.837900 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="registry-server" Nov 24 20:23:36 crc kubenswrapper[5035]: E1124 20:23:36.837930 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="extract-content" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.837942 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="extract-content" Nov 24 20:23:36 crc kubenswrapper[5035]: E1124 20:23:36.837989 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="extract-utilities" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.838005 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="extract-utilities" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.838238 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5729c5ed-8650-41ed-85a6-67d1f9804039" containerName="registry-server" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.839970 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.879072 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6f4sh"] Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.883978 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-utilities\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.884095 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-catalog-content\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.884203 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhqkr\" (UniqueName: \"kubernetes.io/projected/c2948775-c6a0-4658-acfe-af5f77242796-kube-api-access-fhqkr\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.986726 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-catalog-content\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.986882 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhqkr\" (UniqueName: \"kubernetes.io/projected/c2948775-c6a0-4658-acfe-af5f77242796-kube-api-access-fhqkr\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.987115 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-utilities\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.987513 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-catalog-content\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:36 crc kubenswrapper[5035]: I1124 20:23:36.987545 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-utilities\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:37 crc kubenswrapper[5035]: I1124 20:23:37.006754 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhqkr\" (UniqueName: \"kubernetes.io/projected/c2948775-c6a0-4658-acfe-af5f77242796-kube-api-access-fhqkr\") pod \"community-operators-6f4sh\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:37 crc kubenswrapper[5035]: I1124 20:23:37.183611 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:37 crc kubenswrapper[5035]: I1124 20:23:37.771041 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6f4sh"] Nov 24 20:23:37 crc kubenswrapper[5035]: W1124 20:23:37.776981 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2948775_c6a0_4658_acfe_af5f77242796.slice/crio-7a437beb8dbdbfeff846d53b7ea7b8d92437e23640d1e881dcfa33a7f50eac12 WatchSource:0}: Error finding container 7a437beb8dbdbfeff846d53b7ea7b8d92437e23640d1e881dcfa33a7f50eac12: Status 404 returned error can't find the container with id 7a437beb8dbdbfeff846d53b7ea7b8d92437e23640d1e881dcfa33a7f50eac12 Nov 24 20:23:38 crc kubenswrapper[5035]: I1124 20:23:38.746477 5035 generic.go:334] "Generic (PLEG): container finished" podID="c2948775-c6a0-4658-acfe-af5f77242796" containerID="a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9" exitCode=0 Nov 24 20:23:38 crc kubenswrapper[5035]: I1124 20:23:38.746656 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f4sh" event={"ID":"c2948775-c6a0-4658-acfe-af5f77242796","Type":"ContainerDied","Data":"a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9"} Nov 24 20:23:38 crc kubenswrapper[5035]: I1124 20:23:38.746854 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f4sh" event={"ID":"c2948775-c6a0-4658-acfe-af5f77242796","Type":"ContainerStarted","Data":"7a437beb8dbdbfeff846d53b7ea7b8d92437e23640d1e881dcfa33a7f50eac12"} Nov 24 20:23:40 crc kubenswrapper[5035]: I1124 20:23:40.767837 5035 generic.go:334] "Generic (PLEG): container finished" podID="c2948775-c6a0-4658-acfe-af5f77242796" containerID="4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e" exitCode=0 Nov 24 20:23:40 crc kubenswrapper[5035]: I1124 20:23:40.768080 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f4sh" event={"ID":"c2948775-c6a0-4658-acfe-af5f77242796","Type":"ContainerDied","Data":"4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e"} Nov 24 20:23:41 crc kubenswrapper[5035]: I1124 20:23:41.780191 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f4sh" event={"ID":"c2948775-c6a0-4658-acfe-af5f77242796","Type":"ContainerStarted","Data":"fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e"} Nov 24 20:23:41 crc kubenswrapper[5035]: I1124 20:23:41.810736 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6f4sh" podStartSLOduration=3.199908911 podStartE2EDuration="5.810708513s" podCreationTimestamp="2025-11-24 20:23:36 +0000 UTC" firstStartedPulling="2025-11-24 20:23:38.748572341 +0000 UTC m=+4157.271078608" lastFinishedPulling="2025-11-24 20:23:41.359371913 +0000 UTC m=+4159.881878210" observedRunningTime="2025-11-24 20:23:41.804035185 +0000 UTC m=+4160.326541482" watchObservedRunningTime="2025-11-24 20:23:41.810708513 +0000 UTC m=+4160.333214790" Nov 24 20:23:47 crc kubenswrapper[5035]: I1124 20:23:47.184809 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:47 crc kubenswrapper[5035]: I1124 20:23:47.185536 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:47 crc kubenswrapper[5035]: I1124 20:23:47.273570 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:47 crc kubenswrapper[5035]: I1124 20:23:47.923831 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:47 crc kubenswrapper[5035]: I1124 20:23:47.990631 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6f4sh"] Nov 24 20:23:49 crc kubenswrapper[5035]: I1124 20:23:49.869094 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6f4sh" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="registry-server" containerID="cri-o://fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e" gracePeriod=2 Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.464646 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.500069 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-utilities\") pod \"c2948775-c6a0-4658-acfe-af5f77242796\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.500144 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhqkr\" (UniqueName: \"kubernetes.io/projected/c2948775-c6a0-4658-acfe-af5f77242796-kube-api-access-fhqkr\") pod \"c2948775-c6a0-4658-acfe-af5f77242796\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.500275 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-catalog-content\") pod \"c2948775-c6a0-4658-acfe-af5f77242796\" (UID: \"c2948775-c6a0-4658-acfe-af5f77242796\") " Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.503059 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-utilities" (OuterVolumeSpecName: "utilities") pod "c2948775-c6a0-4658-acfe-af5f77242796" (UID: "c2948775-c6a0-4658-acfe-af5f77242796"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.515569 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2948775-c6a0-4658-acfe-af5f77242796-kube-api-access-fhqkr" (OuterVolumeSpecName: "kube-api-access-fhqkr") pod "c2948775-c6a0-4658-acfe-af5f77242796" (UID: "c2948775-c6a0-4658-acfe-af5f77242796"). InnerVolumeSpecName "kube-api-access-fhqkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.602820 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.602857 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhqkr\" (UniqueName: \"kubernetes.io/projected/c2948775-c6a0-4658-acfe-af5f77242796-kube-api-access-fhqkr\") on node \"crc\" DevicePath \"\"" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.882106 5035 generic.go:334] "Generic (PLEG): container finished" podID="c2948775-c6a0-4658-acfe-af5f77242796" containerID="fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e" exitCode=0 Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.882142 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f4sh" event={"ID":"c2948775-c6a0-4658-acfe-af5f77242796","Type":"ContainerDied","Data":"fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e"} Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.882166 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f4sh" event={"ID":"c2948775-c6a0-4658-acfe-af5f77242796","Type":"ContainerDied","Data":"7a437beb8dbdbfeff846d53b7ea7b8d92437e23640d1e881dcfa33a7f50eac12"} Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.882193 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f4sh" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.882193 5035 scope.go:117] "RemoveContainer" containerID="fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.908949 5035 scope.go:117] "RemoveContainer" containerID="4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.934370 5035 scope.go:117] "RemoveContainer" containerID="a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.976438 5035 scope.go:117] "RemoveContainer" containerID="fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e" Nov 24 20:23:50 crc kubenswrapper[5035]: E1124 20:23:50.976853 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e\": container with ID starting with fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e not found: ID does not exist" containerID="fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.976901 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e"} err="failed to get container status \"fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e\": rpc error: code = NotFound desc = could not find container \"fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e\": container with ID starting with fadb84eaf7ad5bf975e67782abc565c62dc0849b22badd892201bfdec138128e not found: ID does not exist" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.976934 5035 scope.go:117] "RemoveContainer" containerID="4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e" Nov 24 20:23:50 crc kubenswrapper[5035]: E1124 20:23:50.977465 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e\": container with ID starting with 4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e not found: ID does not exist" containerID="4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.977507 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e"} err="failed to get container status \"4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e\": rpc error: code = NotFound desc = could not find container \"4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e\": container with ID starting with 4d47a638cce596bc178875583c4ddcc0349449d0495250cdee5b8b59c421f37e not found: ID does not exist" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.977532 5035 scope.go:117] "RemoveContainer" containerID="a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9" Nov 24 20:23:50 crc kubenswrapper[5035]: E1124 20:23:50.977846 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9\": container with ID starting with a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9 not found: ID does not exist" containerID="a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9" Nov 24 20:23:50 crc kubenswrapper[5035]: I1124 20:23:50.977875 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9"} err="failed to get container status \"a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9\": rpc error: code = NotFound desc = could not find container \"a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9\": container with ID starting with a5cbc9427a0bf72d6f0f9d0297503b4486682693384e69d41e9d1f5b989983b9 not found: ID does not exist" Nov 24 20:23:51 crc kubenswrapper[5035]: I1124 20:23:51.547553 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2948775-c6a0-4658-acfe-af5f77242796" (UID: "c2948775-c6a0-4658-acfe-af5f77242796"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:23:51 crc kubenswrapper[5035]: I1124 20:23:51.623990 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2948775-c6a0-4658-acfe-af5f77242796-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:23:51 crc kubenswrapper[5035]: I1124 20:23:51.821108 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6f4sh"] Nov 24 20:23:51 crc kubenswrapper[5035]: I1124 20:23:51.830561 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6f4sh"] Nov 24 20:23:52 crc kubenswrapper[5035]: I1124 20:23:52.217527 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2948775-c6a0-4658-acfe-af5f77242796" path="/var/lib/kubelet/pods/c2948775-c6a0-4658-acfe-af5f77242796/volumes" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.541803 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-db4x7"] Nov 24 20:24:20 crc kubenswrapper[5035]: E1124 20:24:20.542837 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="registry-server" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.542857 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="registry-server" Nov 24 20:24:20 crc kubenswrapper[5035]: E1124 20:24:20.542872 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="extract-content" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.542879 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="extract-content" Nov 24 20:24:20 crc kubenswrapper[5035]: E1124 20:24:20.542910 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="extract-utilities" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.542917 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="extract-utilities" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.543135 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2948775-c6a0-4658-acfe-af5f77242796" containerName="registry-server" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.544828 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.562215 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-db4x7"] Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.712375 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpzsc\" (UniqueName: \"kubernetes.io/projected/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-kube-api-access-qpzsc\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.712915 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-catalog-content\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.712956 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-utilities\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.815114 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpzsc\" (UniqueName: \"kubernetes.io/projected/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-kube-api-access-qpzsc\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.815254 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-catalog-content\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.815282 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-utilities\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.815884 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-utilities\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.815939 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-catalog-content\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.841676 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpzsc\" (UniqueName: \"kubernetes.io/projected/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-kube-api-access-qpzsc\") pod \"redhat-marketplace-db4x7\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:20 crc kubenswrapper[5035]: I1124 20:24:20.880689 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:21 crc kubenswrapper[5035]: I1124 20:24:21.410925 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-db4x7"] Nov 24 20:24:21 crc kubenswrapper[5035]: W1124 20:24:21.419156 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c824d9d_f598_4546_8a4f_72c71fb5cdb9.slice/crio-a1682fff14edb7d63e90883894ad190140c2f19aafd82a632618ae6bf09deb47 WatchSource:0}: Error finding container a1682fff14edb7d63e90883894ad190140c2f19aafd82a632618ae6bf09deb47: Status 404 returned error can't find the container with id a1682fff14edb7d63e90883894ad190140c2f19aafd82a632618ae6bf09deb47 Nov 24 20:24:22 crc kubenswrapper[5035]: I1124 20:24:22.267538 5035 generic.go:334] "Generic (PLEG): container finished" podID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerID="00f534357d026f898378ce3eaa2305bd4852d534a91bd3cfaa0bdb257355047c" exitCode=0 Nov 24 20:24:22 crc kubenswrapper[5035]: I1124 20:24:22.267685 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db4x7" event={"ID":"3c824d9d-f598-4546-8a4f-72c71fb5cdb9","Type":"ContainerDied","Data":"00f534357d026f898378ce3eaa2305bd4852d534a91bd3cfaa0bdb257355047c"} Nov 24 20:24:22 crc kubenswrapper[5035]: I1124 20:24:22.268005 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db4x7" event={"ID":"3c824d9d-f598-4546-8a4f-72c71fb5cdb9","Type":"ContainerStarted","Data":"a1682fff14edb7d63e90883894ad190140c2f19aafd82a632618ae6bf09deb47"} Nov 24 20:24:23 crc kubenswrapper[5035]: I1124 20:24:23.284327 5035 generic.go:334] "Generic (PLEG): container finished" podID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerID="a0605302dbcee12b64026ca246997383fbebea7905713cc06471b3499fa5f8e0" exitCode=0 Nov 24 20:24:23 crc kubenswrapper[5035]: I1124 20:24:23.284443 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db4x7" event={"ID":"3c824d9d-f598-4546-8a4f-72c71fb5cdb9","Type":"ContainerDied","Data":"a0605302dbcee12b64026ca246997383fbebea7905713cc06471b3499fa5f8e0"} Nov 24 20:24:24 crc kubenswrapper[5035]: I1124 20:24:24.293352 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db4x7" event={"ID":"3c824d9d-f598-4546-8a4f-72c71fb5cdb9","Type":"ContainerStarted","Data":"bc30cf97213f8541c99952fdfd9b48286b9abd2b28dbdbddb4551b8731cf61ff"} Nov 24 20:24:24 crc kubenswrapper[5035]: I1124 20:24:24.315383 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-db4x7" podStartSLOduration=2.9311726 podStartE2EDuration="4.315368091s" podCreationTimestamp="2025-11-24 20:24:20 +0000 UTC" firstStartedPulling="2025-11-24 20:24:22.27134775 +0000 UTC m=+4200.793854047" lastFinishedPulling="2025-11-24 20:24:23.655543281 +0000 UTC m=+4202.178049538" observedRunningTime="2025-11-24 20:24:24.312164025 +0000 UTC m=+4202.834670282" watchObservedRunningTime="2025-11-24 20:24:24.315368091 +0000 UTC m=+4202.837874348" Nov 24 20:24:30 crc kubenswrapper[5035]: I1124 20:24:30.881699 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:30 crc kubenswrapper[5035]: I1124 20:24:30.882354 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:30 crc kubenswrapper[5035]: I1124 20:24:30.937747 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:31 crc kubenswrapper[5035]: I1124 20:24:31.588484 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:31 crc kubenswrapper[5035]: I1124 20:24:31.639566 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-db4x7"] Nov 24 20:24:33 crc kubenswrapper[5035]: I1124 20:24:33.409406 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-db4x7" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="registry-server" containerID="cri-o://bc30cf97213f8541c99952fdfd9b48286b9abd2b28dbdbddb4551b8731cf61ff" gracePeriod=2 Nov 24 20:24:34 crc kubenswrapper[5035]: I1124 20:24:34.422924 5035 generic.go:334] "Generic (PLEG): container finished" podID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerID="bc30cf97213f8541c99952fdfd9b48286b9abd2b28dbdbddb4551b8731cf61ff" exitCode=0 Nov 24 20:24:34 crc kubenswrapper[5035]: I1124 20:24:34.422970 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db4x7" event={"ID":"3c824d9d-f598-4546-8a4f-72c71fb5cdb9","Type":"ContainerDied","Data":"bc30cf97213f8541c99952fdfd9b48286b9abd2b28dbdbddb4551b8731cf61ff"} Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.134993 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.229325 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpzsc\" (UniqueName: \"kubernetes.io/projected/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-kube-api-access-qpzsc\") pod \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.229448 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-catalog-content\") pod \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.229704 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-utilities\") pod \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\" (UID: \"3c824d9d-f598-4546-8a4f-72c71fb5cdb9\") " Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.230668 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-utilities" (OuterVolumeSpecName: "utilities") pod "3c824d9d-f598-4546-8a4f-72c71fb5cdb9" (UID: "3c824d9d-f598-4546-8a4f-72c71fb5cdb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.231202 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.238479 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-kube-api-access-qpzsc" (OuterVolumeSpecName: "kube-api-access-qpzsc") pod "3c824d9d-f598-4546-8a4f-72c71fb5cdb9" (UID: "3c824d9d-f598-4546-8a4f-72c71fb5cdb9"). InnerVolumeSpecName "kube-api-access-qpzsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.249019 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c824d9d-f598-4546-8a4f-72c71fb5cdb9" (UID: "3c824d9d-f598-4546-8a4f-72c71fb5cdb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.333701 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpzsc\" (UniqueName: \"kubernetes.io/projected/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-kube-api-access-qpzsc\") on node \"crc\" DevicePath \"\"" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.333737 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c824d9d-f598-4546-8a4f-72c71fb5cdb9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.434538 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db4x7" event={"ID":"3c824d9d-f598-4546-8a4f-72c71fb5cdb9","Type":"ContainerDied","Data":"a1682fff14edb7d63e90883894ad190140c2f19aafd82a632618ae6bf09deb47"} Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.434589 5035 scope.go:117] "RemoveContainer" containerID="bc30cf97213f8541c99952fdfd9b48286b9abd2b28dbdbddb4551b8731cf61ff" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.434755 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db4x7" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.462729 5035 scope.go:117] "RemoveContainer" containerID="a0605302dbcee12b64026ca246997383fbebea7905713cc06471b3499fa5f8e0" Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.474116 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-db4x7"] Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.482261 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-db4x7"] Nov 24 20:24:35 crc kubenswrapper[5035]: I1124 20:24:35.502655 5035 scope.go:117] "RemoveContainer" containerID="00f534357d026f898378ce3eaa2305bd4852d534a91bd3cfaa0bdb257355047c" Nov 24 20:24:36 crc kubenswrapper[5035]: I1124 20:24:36.215185 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" path="/var/lib/kubelet/pods/3c824d9d-f598-4546-8a4f-72c71fb5cdb9/volumes" Nov 24 20:25:15 crc kubenswrapper[5035]: I1124 20:25:15.235110 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:25:15 crc kubenswrapper[5035]: I1124 20:25:15.235876 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:25:45 crc kubenswrapper[5035]: I1124 20:25:45.234924 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:25:45 crc kubenswrapper[5035]: I1124 20:25:45.235651 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:26:15 crc kubenswrapper[5035]: I1124 20:26:15.234378 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:26:15 crc kubenswrapper[5035]: I1124 20:26:15.235074 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:26:15 crc kubenswrapper[5035]: I1124 20:26:15.235146 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:26:15 crc kubenswrapper[5035]: I1124 20:26:15.236280 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:26:15 crc kubenswrapper[5035]: I1124 20:26:15.236417 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" gracePeriod=600 Nov 24 20:26:16 crc kubenswrapper[5035]: I1124 20:26:16.486751 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" exitCode=0 Nov 24 20:26:16 crc kubenswrapper[5035]: I1124 20:26:16.486924 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487"} Nov 24 20:26:16 crc kubenswrapper[5035]: I1124 20:26:16.487231 5035 scope.go:117] "RemoveContainer" containerID="ccddfb6bd9389e4f81715231c8b69c39a4065bea00787ec4778a540a35d107d3" Nov 24 20:26:17 crc kubenswrapper[5035]: E1124 20:26:17.477434 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:26:17 crc kubenswrapper[5035]: I1124 20:26:17.503688 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:26:17 crc kubenswrapper[5035]: E1124 20:26:17.504170 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:26:32 crc kubenswrapper[5035]: I1124 20:26:32.205182 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:26:32 crc kubenswrapper[5035]: E1124 20:26:32.206817 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:26:46 crc kubenswrapper[5035]: I1124 20:26:46.200961 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:26:46 crc kubenswrapper[5035]: E1124 20:26:46.202193 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:27:00 crc kubenswrapper[5035]: I1124 20:27:00.200023 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:27:00 crc kubenswrapper[5035]: E1124 20:27:00.200779 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:27:13 crc kubenswrapper[5035]: I1124 20:27:13.200456 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:27:13 crc kubenswrapper[5035]: E1124 20:27:13.201206 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:27:27 crc kubenswrapper[5035]: I1124 20:27:27.201793 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:27:27 crc kubenswrapper[5035]: E1124 20:27:27.203068 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:27:40 crc kubenswrapper[5035]: I1124 20:27:40.199691 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:27:40 crc kubenswrapper[5035]: E1124 20:27:40.200240 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.761869 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rkdl9"] Nov 24 20:27:44 crc kubenswrapper[5035]: E1124 20:27:44.762933 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="extract-content" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.762954 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="extract-content" Nov 24 20:27:44 crc kubenswrapper[5035]: E1124 20:27:44.762986 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="registry-server" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.762994 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="registry-server" Nov 24 20:27:44 crc kubenswrapper[5035]: E1124 20:27:44.763012 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="extract-utilities" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.763021 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="extract-utilities" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.763243 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c824d9d-f598-4546-8a4f-72c71fb5cdb9" containerName="registry-server" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.764942 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.774461 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkdl9"] Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.867570 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4faf7893-a5a7-4f59-b03c-b2c95642f103-utilities\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.867709 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4faf7893-a5a7-4f59-b03c-b2c95642f103-catalog-content\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.867768 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqnkw\" (UniqueName: \"kubernetes.io/projected/4faf7893-a5a7-4f59-b03c-b2c95642f103-kube-api-access-tqnkw\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.969871 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4faf7893-a5a7-4f59-b03c-b2c95642f103-catalog-content\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.969954 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqnkw\" (UniqueName: \"kubernetes.io/projected/4faf7893-a5a7-4f59-b03c-b2c95642f103-kube-api-access-tqnkw\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.970125 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4faf7893-a5a7-4f59-b03c-b2c95642f103-utilities\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.970476 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4faf7893-a5a7-4f59-b03c-b2c95642f103-catalog-content\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.970748 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4faf7893-a5a7-4f59-b03c-b2c95642f103-utilities\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:44 crc kubenswrapper[5035]: I1124 20:27:44.995024 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqnkw\" (UniqueName: \"kubernetes.io/projected/4faf7893-a5a7-4f59-b03c-b2c95642f103-kube-api-access-tqnkw\") pod \"certified-operators-rkdl9\" (UID: \"4faf7893-a5a7-4f59-b03c-b2c95642f103\") " pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:45 crc kubenswrapper[5035]: I1124 20:27:45.096882 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:45 crc kubenswrapper[5035]: I1124 20:27:45.636457 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkdl9"] Nov 24 20:27:46 crc kubenswrapper[5035]: I1124 20:27:46.442440 5035 generic.go:334] "Generic (PLEG): container finished" podID="4faf7893-a5a7-4f59-b03c-b2c95642f103" containerID="6eb25c4acdb8d9e1ac5698b7d4cab804a6797885d6356af04159e470513c339b" exitCode=0 Nov 24 20:27:46 crc kubenswrapper[5035]: I1124 20:27:46.442578 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkdl9" event={"ID":"4faf7893-a5a7-4f59-b03c-b2c95642f103","Type":"ContainerDied","Data":"6eb25c4acdb8d9e1ac5698b7d4cab804a6797885d6356af04159e470513c339b"} Nov 24 20:27:46 crc kubenswrapper[5035]: I1124 20:27:46.442917 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkdl9" event={"ID":"4faf7893-a5a7-4f59-b03c-b2c95642f103","Type":"ContainerStarted","Data":"b5c77ad6ac88ce43cae219ba1afe9f741181dea263003c2a174f1e660b0f500a"} Nov 24 20:27:46 crc kubenswrapper[5035]: I1124 20:27:46.446732 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:27:52 crc kubenswrapper[5035]: I1124 20:27:52.208872 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:27:52 crc kubenswrapper[5035]: E1124 20:27:52.209851 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:27:52 crc kubenswrapper[5035]: I1124 20:27:52.501229 5035 generic.go:334] "Generic (PLEG): container finished" podID="4faf7893-a5a7-4f59-b03c-b2c95642f103" containerID="302a21b23723aa9de5f4677f0fde3bc66f15588c5b6659eae6ba2f9bcf6a5c27" exitCode=0 Nov 24 20:27:52 crc kubenswrapper[5035]: I1124 20:27:52.501270 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkdl9" event={"ID":"4faf7893-a5a7-4f59-b03c-b2c95642f103","Type":"ContainerDied","Data":"302a21b23723aa9de5f4677f0fde3bc66f15588c5b6659eae6ba2f9bcf6a5c27"} Nov 24 20:27:54 crc kubenswrapper[5035]: I1124 20:27:54.523374 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkdl9" event={"ID":"4faf7893-a5a7-4f59-b03c-b2c95642f103","Type":"ContainerStarted","Data":"bfb142db87b3f9c0c0cc2d8e14fc9f3b3ceddd8d9819a0ce3195d9fffaca5832"} Nov 24 20:27:54 crc kubenswrapper[5035]: I1124 20:27:54.552469 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rkdl9" podStartSLOduration=3.1445977210000002 podStartE2EDuration="10.552443587s" podCreationTimestamp="2025-11-24 20:27:44 +0000 UTC" firstStartedPulling="2025-11-24 20:27:46.446504774 +0000 UTC m=+4404.969011031" lastFinishedPulling="2025-11-24 20:27:53.85435063 +0000 UTC m=+4412.376856897" observedRunningTime="2025-11-24 20:27:54.549055676 +0000 UTC m=+4413.071561953" watchObservedRunningTime="2025-11-24 20:27:54.552443587 +0000 UTC m=+4413.074949854" Nov 24 20:27:55 crc kubenswrapper[5035]: I1124 20:27:55.097440 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:55 crc kubenswrapper[5035]: I1124 20:27:55.097647 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:27:56 crc kubenswrapper[5035]: I1124 20:27:56.166520 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rkdl9" podUID="4faf7893-a5a7-4f59-b03c-b2c95642f103" containerName="registry-server" probeResult="failure" output=< Nov 24 20:27:56 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:27:56 crc kubenswrapper[5035]: > Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.174181 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.200742 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:28:05 crc kubenswrapper[5035]: E1124 20:28:05.201409 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.243507 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rkdl9" Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.352778 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkdl9"] Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.437643 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k54vj"] Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.437893 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k54vj" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="registry-server" containerID="cri-o://3a3fdb05890dc52f33afdd871678f92398a79d895e3fba8552600c255db7540f" gracePeriod=2 Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.632610 5035 generic.go:334] "Generic (PLEG): container finished" podID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerID="3a3fdb05890dc52f33afdd871678f92398a79d895e3fba8552600c255db7540f" exitCode=0 Nov 24 20:28:05 crc kubenswrapper[5035]: I1124 20:28:05.632812 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k54vj" event={"ID":"2295440b-2755-4332-8d15-5e2afdeefc9e","Type":"ContainerDied","Data":"3a3fdb05890dc52f33afdd871678f92398a79d895e3fba8552600c255db7540f"} Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.091060 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.233205 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-utilities\") pod \"2295440b-2755-4332-8d15-5e2afdeefc9e\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.233371 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-catalog-content\") pod \"2295440b-2755-4332-8d15-5e2afdeefc9e\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.233478 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fq8h\" (UniqueName: \"kubernetes.io/projected/2295440b-2755-4332-8d15-5e2afdeefc9e-kube-api-access-6fq8h\") pod \"2295440b-2755-4332-8d15-5e2afdeefc9e\" (UID: \"2295440b-2755-4332-8d15-5e2afdeefc9e\") " Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.233683 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-utilities" (OuterVolumeSpecName: "utilities") pod "2295440b-2755-4332-8d15-5e2afdeefc9e" (UID: "2295440b-2755-4332-8d15-5e2afdeefc9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.234247 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.270482 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2295440b-2755-4332-8d15-5e2afdeefc9e-kube-api-access-6fq8h" (OuterVolumeSpecName: "kube-api-access-6fq8h") pod "2295440b-2755-4332-8d15-5e2afdeefc9e" (UID: "2295440b-2755-4332-8d15-5e2afdeefc9e"). InnerVolumeSpecName "kube-api-access-6fq8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.275376 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2295440b-2755-4332-8d15-5e2afdeefc9e" (UID: "2295440b-2755-4332-8d15-5e2afdeefc9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.336434 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2295440b-2755-4332-8d15-5e2afdeefc9e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.336470 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fq8h\" (UniqueName: \"kubernetes.io/projected/2295440b-2755-4332-8d15-5e2afdeefc9e-kube-api-access-6fq8h\") on node \"crc\" DevicePath \"\"" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.650183 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k54vj" event={"ID":"2295440b-2755-4332-8d15-5e2afdeefc9e","Type":"ContainerDied","Data":"7f7684801806a4c8b3c59e431012d33d8547944b46089c64de6ca0d6433581b4"} Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.650359 5035 scope.go:117] "RemoveContainer" containerID="3a3fdb05890dc52f33afdd871678f92398a79d895e3fba8552600c255db7540f" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.650242 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k54vj" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.701500 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k54vj"] Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.712408 5035 scope.go:117] "RemoveContainer" containerID="52023767e730293ce283ce003facee4708e99e2a3a5bf21d941e9bf7ef9d2036" Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.715012 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k54vj"] Nov 24 20:28:06 crc kubenswrapper[5035]: I1124 20:28:06.762161 5035 scope.go:117] "RemoveContainer" containerID="5d9dd9e9903da1ec31180ce4be78fc8f2a2d1a39f733518951548663af1dc237" Nov 24 20:28:08 crc kubenswrapper[5035]: I1124 20:28:08.218252 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" path="/var/lib/kubelet/pods/2295440b-2755-4332-8d15-5e2afdeefc9e/volumes" Nov 24 20:28:19 crc kubenswrapper[5035]: I1124 20:28:19.201459 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:28:19 crc kubenswrapper[5035]: E1124 20:28:19.202544 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:28:32 crc kubenswrapper[5035]: I1124 20:28:32.216462 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:28:32 crc kubenswrapper[5035]: E1124 20:28:32.217747 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:28:46 crc kubenswrapper[5035]: I1124 20:28:46.200400 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:28:46 crc kubenswrapper[5035]: E1124 20:28:46.201566 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:29:00 crc kubenswrapper[5035]: I1124 20:29:00.199958 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:29:00 crc kubenswrapper[5035]: E1124 20:29:00.202186 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:29:13 crc kubenswrapper[5035]: I1124 20:29:13.200461 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:29:13 crc kubenswrapper[5035]: E1124 20:29:13.201354 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:29:24 crc kubenswrapper[5035]: I1124 20:29:24.200408 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:29:24 crc kubenswrapper[5035]: E1124 20:29:24.201712 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:29:37 crc kubenswrapper[5035]: I1124 20:29:37.201221 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:29:37 crc kubenswrapper[5035]: E1124 20:29:37.202261 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:29:50 crc kubenswrapper[5035]: I1124 20:29:50.200191 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:29:50 crc kubenswrapper[5035]: E1124 20:29:50.201002 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.146890 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t"] Nov 24 20:30:00 crc kubenswrapper[5035]: E1124 20:30:00.147815 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="extract-utilities" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.147830 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="extract-utilities" Nov 24 20:30:00 crc kubenswrapper[5035]: E1124 20:30:00.147856 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="extract-content" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.147861 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="extract-content" Nov 24 20:30:00 crc kubenswrapper[5035]: E1124 20:30:00.147875 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="registry-server" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.147881 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="registry-server" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.148077 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2295440b-2755-4332-8d15-5e2afdeefc9e" containerName="registry-server" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.148750 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.150954 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.151694 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.162430 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t"] Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.242563 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn5rw\" (UniqueName: \"kubernetes.io/projected/e9d1d182-8018-4fe4-804f-aa198fdf4812-kube-api-access-pn5rw\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.242612 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9d1d182-8018-4fe4-804f-aa198fdf4812-config-volume\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.242665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9d1d182-8018-4fe4-804f-aa198fdf4812-secret-volume\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.344719 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn5rw\" (UniqueName: \"kubernetes.io/projected/e9d1d182-8018-4fe4-804f-aa198fdf4812-kube-api-access-pn5rw\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.345029 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9d1d182-8018-4fe4-804f-aa198fdf4812-config-volume\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.345246 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9d1d182-8018-4fe4-804f-aa198fdf4812-secret-volume\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.346066 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9d1d182-8018-4fe4-804f-aa198fdf4812-config-volume\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.350415 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9d1d182-8018-4fe4-804f-aa198fdf4812-secret-volume\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.363325 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn5rw\" (UniqueName: \"kubernetes.io/projected/e9d1d182-8018-4fe4-804f-aa198fdf4812-kube-api-access-pn5rw\") pod \"collect-profiles-29400270-qpb4t\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:00 crc kubenswrapper[5035]: I1124 20:30:00.467178 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:01 crc kubenswrapper[5035]: I1124 20:30:01.002738 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t"] Nov 24 20:30:01 crc kubenswrapper[5035]: I1124 20:30:01.844561 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" event={"ID":"e9d1d182-8018-4fe4-804f-aa198fdf4812","Type":"ContainerStarted","Data":"6f579b2454f0b077d89c1ab3db50d1ba0358791295b7274b4c22594ab613587c"} Nov 24 20:30:01 crc kubenswrapper[5035]: I1124 20:30:01.844863 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" event={"ID":"e9d1d182-8018-4fe4-804f-aa198fdf4812","Type":"ContainerStarted","Data":"3b1a3b64eb1fbcc1e07b20d8d8f645e7d110ab7921fc4fca14b7bf8ee7018c15"} Nov 24 20:30:01 crc kubenswrapper[5035]: I1124 20:30:01.861194 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" podStartSLOduration=1.8611749610000001 podStartE2EDuration="1.861174961s" podCreationTimestamp="2025-11-24 20:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:30:01.85855002 +0000 UTC m=+4540.381056277" watchObservedRunningTime="2025-11-24 20:30:01.861174961 +0000 UTC m=+4540.383681218" Nov 24 20:30:02 crc kubenswrapper[5035]: I1124 20:30:02.852478 5035 generic.go:334] "Generic (PLEG): container finished" podID="e9d1d182-8018-4fe4-804f-aa198fdf4812" containerID="6f579b2454f0b077d89c1ab3db50d1ba0358791295b7274b4c22594ab613587c" exitCode=0 Nov 24 20:30:02 crc kubenswrapper[5035]: I1124 20:30:02.852577 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" event={"ID":"e9d1d182-8018-4fe4-804f-aa198fdf4812","Type":"ContainerDied","Data":"6f579b2454f0b077d89c1ab3db50d1ba0358791295b7274b4c22594ab613587c"} Nov 24 20:30:03 crc kubenswrapper[5035]: I1124 20:30:03.200609 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:30:03 crc kubenswrapper[5035]: E1124 20:30:03.201238 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.657648 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.753629 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn5rw\" (UniqueName: \"kubernetes.io/projected/e9d1d182-8018-4fe4-804f-aa198fdf4812-kube-api-access-pn5rw\") pod \"e9d1d182-8018-4fe4-804f-aa198fdf4812\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.753712 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9d1d182-8018-4fe4-804f-aa198fdf4812-secret-volume\") pod \"e9d1d182-8018-4fe4-804f-aa198fdf4812\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.753832 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9d1d182-8018-4fe4-804f-aa198fdf4812-config-volume\") pod \"e9d1d182-8018-4fe4-804f-aa198fdf4812\" (UID: \"e9d1d182-8018-4fe4-804f-aa198fdf4812\") " Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.755124 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9d1d182-8018-4fe4-804f-aa198fdf4812-config-volume" (OuterVolumeSpecName: "config-volume") pod "e9d1d182-8018-4fe4-804f-aa198fdf4812" (UID: "e9d1d182-8018-4fe4-804f-aa198fdf4812"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.774847 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d1d182-8018-4fe4-804f-aa198fdf4812-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e9d1d182-8018-4fe4-804f-aa198fdf4812" (UID: "e9d1d182-8018-4fe4-804f-aa198fdf4812"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.782725 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d1d182-8018-4fe4-804f-aa198fdf4812-kube-api-access-pn5rw" (OuterVolumeSpecName: "kube-api-access-pn5rw") pod "e9d1d182-8018-4fe4-804f-aa198fdf4812" (UID: "e9d1d182-8018-4fe4-804f-aa198fdf4812"). InnerVolumeSpecName "kube-api-access-pn5rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.856537 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn5rw\" (UniqueName: \"kubernetes.io/projected/e9d1d182-8018-4fe4-804f-aa198fdf4812-kube-api-access-pn5rw\") on node \"crc\" DevicePath \"\"" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.856578 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9d1d182-8018-4fe4-804f-aa198fdf4812-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.856590 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9d1d182-8018-4fe4-804f-aa198fdf4812-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.898111 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" event={"ID":"e9d1d182-8018-4fe4-804f-aa198fdf4812","Type":"ContainerDied","Data":"3b1a3b64eb1fbcc1e07b20d8d8f645e7d110ab7921fc4fca14b7bf8ee7018c15"} Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.898160 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b1a3b64eb1fbcc1e07b20d8d8f645e7d110ab7921fc4fca14b7bf8ee7018c15" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.898220 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t" Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.941063 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx"] Nov 24 20:30:04 crc kubenswrapper[5035]: I1124 20:30:04.947867 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400225-v6cvx"] Nov 24 20:30:05 crc kubenswrapper[5035]: E1124 20:30:05.122505 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9d1d182_8018_4fe4_804f_aa198fdf4812.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9d1d182_8018_4fe4_804f_aa198fdf4812.slice/crio-3b1a3b64eb1fbcc1e07b20d8d8f645e7d110ab7921fc4fca14b7bf8ee7018c15\": RecentStats: unable to find data in memory cache]" Nov 24 20:30:06 crc kubenswrapper[5035]: I1124 20:30:06.214342 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b2065a7-34e8-4aa0-ab18-60c002c5fbc7" path="/var/lib/kubelet/pods/3b2065a7-34e8-4aa0-ab18-60c002c5fbc7/volumes" Nov 24 20:30:14 crc kubenswrapper[5035]: I1124 20:30:14.200395 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:30:14 crc kubenswrapper[5035]: E1124 20:30:14.201738 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:30:29 crc kubenswrapper[5035]: I1124 20:30:29.201102 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:30:29 crc kubenswrapper[5035]: E1124 20:30:29.201846 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:30:31 crc kubenswrapper[5035]: I1124 20:30:31.667086 5035 scope.go:117] "RemoveContainer" containerID="298a61232d0326191769c34ceede8538573bdc760d0a8398c6cc9760d7f349a3" Nov 24 20:30:41 crc kubenswrapper[5035]: I1124 20:30:41.200477 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:30:41 crc kubenswrapper[5035]: E1124 20:30:41.201285 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:30:52 crc kubenswrapper[5035]: I1124 20:30:52.216270 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:30:52 crc kubenswrapper[5035]: E1124 20:30:52.217428 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:31:07 crc kubenswrapper[5035]: I1124 20:31:07.200481 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:31:07 crc kubenswrapper[5035]: E1124 20:31:07.201261 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.381802 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gdls6"] Nov 24 20:31:08 crc kubenswrapper[5035]: E1124 20:31:08.383523 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d1d182-8018-4fe4-804f-aa198fdf4812" containerName="collect-profiles" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.383633 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d1d182-8018-4fe4-804f-aa198fdf4812" containerName="collect-profiles" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.383869 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d1d182-8018-4fe4-804f-aa198fdf4812" containerName="collect-profiles" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.385641 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.399638 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdls6"] Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.428016 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-utilities\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.428336 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-catalog-content\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.428414 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h2dt\" (UniqueName: \"kubernetes.io/projected/ad9d720e-ee48-4653-b01e-cda40ef52ed7-kube-api-access-8h2dt\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.529376 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-utilities\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.529582 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-catalog-content\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.529631 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h2dt\" (UniqueName: \"kubernetes.io/projected/ad9d720e-ee48-4653-b01e-cda40ef52ed7-kube-api-access-8h2dt\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.530085 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-utilities\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.530088 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-catalog-content\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.556213 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h2dt\" (UniqueName: \"kubernetes.io/projected/ad9d720e-ee48-4653-b01e-cda40ef52ed7-kube-api-access-8h2dt\") pod \"redhat-operators-gdls6\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:08 crc kubenswrapper[5035]: I1124 20:31:08.708615 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:09 crc kubenswrapper[5035]: I1124 20:31:09.175294 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdls6"] Nov 24 20:31:09 crc kubenswrapper[5035]: I1124 20:31:09.476414 5035 generic.go:334] "Generic (PLEG): container finished" podID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerID="1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a" exitCode=0 Nov 24 20:31:09 crc kubenswrapper[5035]: I1124 20:31:09.476471 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdls6" event={"ID":"ad9d720e-ee48-4653-b01e-cda40ef52ed7","Type":"ContainerDied","Data":"1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a"} Nov 24 20:31:09 crc kubenswrapper[5035]: I1124 20:31:09.477649 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdls6" event={"ID":"ad9d720e-ee48-4653-b01e-cda40ef52ed7","Type":"ContainerStarted","Data":"0a27045d5f126b70df719185dfbaeefa07063fc1feb9bbc791b79a2df4b1e776"} Nov 24 20:31:11 crc kubenswrapper[5035]: I1124 20:31:11.510148 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdls6" event={"ID":"ad9d720e-ee48-4653-b01e-cda40ef52ed7","Type":"ContainerStarted","Data":"2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a"} Nov 24 20:31:17 crc kubenswrapper[5035]: I1124 20:31:17.568670 5035 generic.go:334] "Generic (PLEG): container finished" podID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerID="2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a" exitCode=0 Nov 24 20:31:17 crc kubenswrapper[5035]: I1124 20:31:17.568757 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdls6" event={"ID":"ad9d720e-ee48-4653-b01e-cda40ef52ed7","Type":"ContainerDied","Data":"2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a"} Nov 24 20:31:18 crc kubenswrapper[5035]: I1124 20:31:18.580806 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdls6" event={"ID":"ad9d720e-ee48-4653-b01e-cda40ef52ed7","Type":"ContainerStarted","Data":"0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173"} Nov 24 20:31:18 crc kubenswrapper[5035]: I1124 20:31:18.597535 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gdls6" podStartSLOduration=2.110208244 podStartE2EDuration="10.597516391s" podCreationTimestamp="2025-11-24 20:31:08 +0000 UTC" firstStartedPulling="2025-11-24 20:31:09.478592493 +0000 UTC m=+4608.001098750" lastFinishedPulling="2025-11-24 20:31:17.96590064 +0000 UTC m=+4616.488406897" observedRunningTime="2025-11-24 20:31:18.596598766 +0000 UTC m=+4617.119105063" watchObservedRunningTime="2025-11-24 20:31:18.597516391 +0000 UTC m=+4617.120022658" Nov 24 20:31:18 crc kubenswrapper[5035]: I1124 20:31:18.709940 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:18 crc kubenswrapper[5035]: I1124 20:31:18.710097 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:19 crc kubenswrapper[5035]: I1124 20:31:19.200802 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:31:19 crc kubenswrapper[5035]: I1124 20:31:19.596550 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"34dfb1c21f73e6324b3482dfbbc8fa0d3090574d28fc0e5aca6749b6c9dd09a2"} Nov 24 20:31:19 crc kubenswrapper[5035]: I1124 20:31:19.765890 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdls6" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="registry-server" probeResult="failure" output=< Nov 24 20:31:19 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:31:19 crc kubenswrapper[5035]: > Nov 24 20:31:29 crc kubenswrapper[5035]: I1124 20:31:29.756447 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdls6" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="registry-server" probeResult="failure" output=< Nov 24 20:31:29 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:31:29 crc kubenswrapper[5035]: > Nov 24 20:31:38 crc kubenswrapper[5035]: I1124 20:31:38.758483 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:38 crc kubenswrapper[5035]: I1124 20:31:38.823388 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:39 crc kubenswrapper[5035]: I1124 20:31:39.584708 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdls6"] Nov 24 20:31:39 crc kubenswrapper[5035]: I1124 20:31:39.805081 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gdls6" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="registry-server" containerID="cri-o://0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173" gracePeriod=2 Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.414990 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.491139 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-catalog-content\") pod \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.491197 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h2dt\" (UniqueName: \"kubernetes.io/projected/ad9d720e-ee48-4653-b01e-cda40ef52ed7-kube-api-access-8h2dt\") pod \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.491344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-utilities\") pod \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\" (UID: \"ad9d720e-ee48-4653-b01e-cda40ef52ed7\") " Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.492604 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-utilities" (OuterVolumeSpecName: "utilities") pod "ad9d720e-ee48-4653-b01e-cda40ef52ed7" (UID: "ad9d720e-ee48-4653-b01e-cda40ef52ed7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.497637 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad9d720e-ee48-4653-b01e-cda40ef52ed7-kube-api-access-8h2dt" (OuterVolumeSpecName: "kube-api-access-8h2dt") pod "ad9d720e-ee48-4653-b01e-cda40ef52ed7" (UID: "ad9d720e-ee48-4653-b01e-cda40ef52ed7"). InnerVolumeSpecName "kube-api-access-8h2dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.593376 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h2dt\" (UniqueName: \"kubernetes.io/projected/ad9d720e-ee48-4653-b01e-cda40ef52ed7-kube-api-access-8h2dt\") on node \"crc\" DevicePath \"\"" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.593410 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.596089 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad9d720e-ee48-4653-b01e-cda40ef52ed7" (UID: "ad9d720e-ee48-4653-b01e-cda40ef52ed7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.695220 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad9d720e-ee48-4653-b01e-cda40ef52ed7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.818109 5035 generic.go:334] "Generic (PLEG): container finished" podID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerID="0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173" exitCode=0 Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.818197 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdls6" event={"ID":"ad9d720e-ee48-4653-b01e-cda40ef52ed7","Type":"ContainerDied","Data":"0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173"} Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.818311 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdls6" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.818346 5035 scope.go:117] "RemoveContainer" containerID="0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.818327 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdls6" event={"ID":"ad9d720e-ee48-4653-b01e-cda40ef52ed7","Type":"ContainerDied","Data":"0a27045d5f126b70df719185dfbaeefa07063fc1feb9bbc791b79a2df4b1e776"} Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.852451 5035 scope.go:117] "RemoveContainer" containerID="2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.867430 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdls6"] Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.876794 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gdls6"] Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.879816 5035 scope.go:117] "RemoveContainer" containerID="1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.926273 5035 scope.go:117] "RemoveContainer" containerID="0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173" Nov 24 20:31:40 crc kubenswrapper[5035]: E1124 20:31:40.927010 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173\": container with ID starting with 0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173 not found: ID does not exist" containerID="0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.927174 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173"} err="failed to get container status \"0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173\": rpc error: code = NotFound desc = could not find container \"0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173\": container with ID starting with 0d922379c905b671c0b707368c2fe6c70378c80c3475b662004b393697b06173 not found: ID does not exist" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.927353 5035 scope.go:117] "RemoveContainer" containerID="2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a" Nov 24 20:31:40 crc kubenswrapper[5035]: E1124 20:31:40.928081 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a\": container with ID starting with 2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a not found: ID does not exist" containerID="2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.928126 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a"} err="failed to get container status \"2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a\": rpc error: code = NotFound desc = could not find container \"2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a\": container with ID starting with 2c385dc4004c3243ad7e0b74b61cce9bf1678c95369919535c96afe6ec91a65a not found: ID does not exist" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.928163 5035 scope.go:117] "RemoveContainer" containerID="1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a" Nov 24 20:31:40 crc kubenswrapper[5035]: E1124 20:31:40.928630 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a\": container with ID starting with 1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a not found: ID does not exist" containerID="1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a" Nov 24 20:31:40 crc kubenswrapper[5035]: I1124 20:31:40.928661 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a"} err="failed to get container status \"1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a\": rpc error: code = NotFound desc = could not find container \"1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a\": container with ID starting with 1a004553eea79d1f0275ab7380169ce2ef9b737b7da46033bc5ae6928ca36d2a not found: ID does not exist" Nov 24 20:31:42 crc kubenswrapper[5035]: I1124 20:31:42.222936 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" path="/var/lib/kubelet/pods/ad9d720e-ee48-4653-b01e-cda40ef52ed7/volumes" Nov 24 20:33:45 crc kubenswrapper[5035]: I1124 20:33:45.234813 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:33:45 crc kubenswrapper[5035]: I1124 20:33:45.235485 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:34:15 crc kubenswrapper[5035]: I1124 20:34:15.233843 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:34:15 crc kubenswrapper[5035]: I1124 20:34:15.234479 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:34:45 crc kubenswrapper[5035]: I1124 20:34:45.234187 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:34:45 crc kubenswrapper[5035]: I1124 20:34:45.234658 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:34:45 crc kubenswrapper[5035]: I1124 20:34:45.234705 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:34:45 crc kubenswrapper[5035]: I1124 20:34:45.235476 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34dfb1c21f73e6324b3482dfbbc8fa0d3090574d28fc0e5aca6749b6c9dd09a2"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:34:45 crc kubenswrapper[5035]: I1124 20:34:45.235545 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://34dfb1c21f73e6324b3482dfbbc8fa0d3090574d28fc0e5aca6749b6c9dd09a2" gracePeriod=600 Nov 24 20:34:46 crc kubenswrapper[5035]: I1124 20:34:46.206128 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="34dfb1c21f73e6324b3482dfbbc8fa0d3090574d28fc0e5aca6749b6c9dd09a2" exitCode=0 Nov 24 20:34:46 crc kubenswrapper[5035]: I1124 20:34:46.213586 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"34dfb1c21f73e6324b3482dfbbc8fa0d3090574d28fc0e5aca6749b6c9dd09a2"} Nov 24 20:34:46 crc kubenswrapper[5035]: I1124 20:34:46.213639 5035 scope.go:117] "RemoveContainer" containerID="a61b65464375d565de202d31a0e850abf919f65911acb6594b0c67ef875a6487" Nov 24 20:34:47 crc kubenswrapper[5035]: I1124 20:34:47.217872 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a"} Nov 24 20:37:15 crc kubenswrapper[5035]: I1124 20:37:15.234817 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:37:15 crc kubenswrapper[5035]: I1124 20:37:15.235608 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:37:45 crc kubenswrapper[5035]: I1124 20:37:45.234305 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:37:45 crc kubenswrapper[5035]: I1124 20:37:45.234835 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.234282 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.235254 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.235320 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.236358 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.236408 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" gracePeriod=600 Nov 24 20:38:15 crc kubenswrapper[5035]: E1124 20:38:15.526400 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.636658 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" exitCode=0 Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.636725 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a"} Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.636761 5035 scope.go:117] "RemoveContainer" containerID="34dfb1c21f73e6324b3482dfbbc8fa0d3090574d28fc0e5aca6749b6c9dd09a2" Nov 24 20:38:15 crc kubenswrapper[5035]: I1124 20:38:15.637736 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:38:15 crc kubenswrapper[5035]: E1124 20:38:15.638080 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:38:28 crc kubenswrapper[5035]: I1124 20:38:28.200413 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:38:28 crc kubenswrapper[5035]: E1124 20:38:28.201202 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:38:41 crc kubenswrapper[5035]: I1124 20:38:41.200997 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:38:41 crc kubenswrapper[5035]: E1124 20:38:41.201796 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:38:52 crc kubenswrapper[5035]: I1124 20:38:52.207691 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:38:52 crc kubenswrapper[5035]: E1124 20:38:52.208736 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:39:04 crc kubenswrapper[5035]: I1124 20:39:04.200823 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:39:04 crc kubenswrapper[5035]: E1124 20:39:04.203007 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:39:17 crc kubenswrapper[5035]: I1124 20:39:17.199819 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:39:17 crc kubenswrapper[5035]: E1124 20:39:17.200630 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:39:28 crc kubenswrapper[5035]: I1124 20:39:28.200337 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:39:28 crc kubenswrapper[5035]: E1124 20:39:28.201160 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:39:39 crc kubenswrapper[5035]: I1124 20:39:39.200317 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:39:39 crc kubenswrapper[5035]: E1124 20:39:39.201162 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:39:51 crc kubenswrapper[5035]: I1124 20:39:51.200674 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:39:51 crc kubenswrapper[5035]: E1124 20:39:51.201657 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:40:02 crc kubenswrapper[5035]: I1124 20:40:02.205862 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:40:02 crc kubenswrapper[5035]: E1124 20:40:02.206515 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:40:15 crc kubenswrapper[5035]: I1124 20:40:15.200023 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:40:15 crc kubenswrapper[5035]: E1124 20:40:15.201049 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:40:28 crc kubenswrapper[5035]: I1124 20:40:28.200198 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:40:28 crc kubenswrapper[5035]: E1124 20:40:28.201913 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:40:41 crc kubenswrapper[5035]: I1124 20:40:41.200718 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:40:41 crc kubenswrapper[5035]: E1124 20:40:41.201659 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:40:52 crc kubenswrapper[5035]: I1124 20:40:52.200760 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:40:52 crc kubenswrapper[5035]: E1124 20:40:52.203039 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.200465 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:41:03 crc kubenswrapper[5035]: E1124 20:41:03.201289 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.408080 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-92qth"] Nov 24 20:41:03 crc kubenswrapper[5035]: E1124 20:41:03.408620 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="registry-server" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.408642 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="registry-server" Nov 24 20:41:03 crc kubenswrapper[5035]: E1124 20:41:03.408661 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="extract-content" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.408669 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="extract-content" Nov 24 20:41:03 crc kubenswrapper[5035]: E1124 20:41:03.408690 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="extract-utilities" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.408698 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="extract-utilities" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.408921 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9d720e-ee48-4653-b01e-cda40ef52ed7" containerName="registry-server" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.410582 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.422741 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-92qth"] Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.535740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhjbk\" (UniqueName: \"kubernetes.io/projected/66eff236-3b10-4ae4-8afd-8e4257ac72de-kube-api-access-zhjbk\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.536084 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-utilities\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.536343 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-catalog-content\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.598054 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sml2k"] Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.602657 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.639370 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-utilities\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.639743 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-catalog-content\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.639978 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-utilities\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.640223 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-catalog-content\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.639948 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhjbk\" (UniqueName: \"kubernetes.io/projected/66eff236-3b10-4ae4-8afd-8e4257ac72de-kube-api-access-zhjbk\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.657137 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sml2k"] Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.687531 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhjbk\" (UniqueName: \"kubernetes.io/projected/66eff236-3b10-4ae4-8afd-8e4257ac72de-kube-api-access-zhjbk\") pod \"certified-operators-92qth\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.742605 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.744036 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-utilities\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.744118 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-catalog-content\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.744160 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tcgv\" (UniqueName: \"kubernetes.io/projected/5962b0dc-5936-455a-8731-12aeabbd0a72-kube-api-access-2tcgv\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.845732 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-utilities\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.845831 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-catalog-content\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.845875 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tcgv\" (UniqueName: \"kubernetes.io/projected/5962b0dc-5936-455a-8731-12aeabbd0a72-kube-api-access-2tcgv\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.847058 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-catalog-content\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.847066 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-utilities\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.867136 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tcgv\" (UniqueName: \"kubernetes.io/projected/5962b0dc-5936-455a-8731-12aeabbd0a72-kube-api-access-2tcgv\") pod \"community-operators-sml2k\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:03 crc kubenswrapper[5035]: I1124 20:41:03.934882 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:04 crc kubenswrapper[5035]: I1124 20:41:04.415271 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-92qth"] Nov 24 20:41:04 crc kubenswrapper[5035]: I1124 20:41:04.654652 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sml2k"] Nov 24 20:41:04 crc kubenswrapper[5035]: I1124 20:41:04.855239 5035 generic.go:334] "Generic (PLEG): container finished" podID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerID="9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4" exitCode=0 Nov 24 20:41:04 crc kubenswrapper[5035]: I1124 20:41:04.855597 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92qth" event={"ID":"66eff236-3b10-4ae4-8afd-8e4257ac72de","Type":"ContainerDied","Data":"9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4"} Nov 24 20:41:04 crc kubenswrapper[5035]: I1124 20:41:04.855650 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92qth" event={"ID":"66eff236-3b10-4ae4-8afd-8e4257ac72de","Type":"ContainerStarted","Data":"9eff37a834b75132e935a1c6f2a1c6b72ec8534323087a52ab3e3b8bc5b958c4"} Nov 24 20:41:04 crc kubenswrapper[5035]: I1124 20:41:04.857115 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:41:04 crc kubenswrapper[5035]: I1124 20:41:04.857709 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml2k" event={"ID":"5962b0dc-5936-455a-8731-12aeabbd0a72","Type":"ContainerStarted","Data":"7c329a53fa99f80ef6f02e30c539151babafbbb377eda08427a3ea469d529893"} Nov 24 20:41:05 crc kubenswrapper[5035]: I1124 20:41:05.871212 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92qth" event={"ID":"66eff236-3b10-4ae4-8afd-8e4257ac72de","Type":"ContainerStarted","Data":"ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff"} Nov 24 20:41:05 crc kubenswrapper[5035]: I1124 20:41:05.872693 5035 generic.go:334] "Generic (PLEG): container finished" podID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerID="fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253" exitCode=0 Nov 24 20:41:05 crc kubenswrapper[5035]: I1124 20:41:05.872745 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml2k" event={"ID":"5962b0dc-5936-455a-8731-12aeabbd0a72","Type":"ContainerDied","Data":"fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253"} Nov 24 20:41:07 crc kubenswrapper[5035]: I1124 20:41:07.907887 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml2k" event={"ID":"5962b0dc-5936-455a-8731-12aeabbd0a72","Type":"ContainerStarted","Data":"6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24"} Nov 24 20:41:07 crc kubenswrapper[5035]: I1124 20:41:07.914343 5035 generic.go:334] "Generic (PLEG): container finished" podID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerID="ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff" exitCode=0 Nov 24 20:41:07 crc kubenswrapper[5035]: I1124 20:41:07.914399 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92qth" event={"ID":"66eff236-3b10-4ae4-8afd-8e4257ac72de","Type":"ContainerDied","Data":"ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff"} Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.224106 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lcvh9"] Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.226888 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lcvh9"] Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.227034 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.251272 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59jcf\" (UniqueName: \"kubernetes.io/projected/69d32786-a0f3-4285-be95-d3c2264cd8d1-kube-api-access-59jcf\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.251388 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-catalog-content\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.251475 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-utilities\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.352751 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-catalog-content\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.352849 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-utilities\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.352960 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59jcf\" (UniqueName: \"kubernetes.io/projected/69d32786-a0f3-4285-be95-d3c2264cd8d1-kube-api-access-59jcf\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.354759 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-catalog-content\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.354825 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-utilities\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.536943 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59jcf\" (UniqueName: \"kubernetes.io/projected/69d32786-a0f3-4285-be95-d3c2264cd8d1-kube-api-access-59jcf\") pod \"redhat-marketplace-lcvh9\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:08 crc kubenswrapper[5035]: I1124 20:41:08.552664 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.138744 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lcvh9"] Nov 24 20:41:09 crc kubenswrapper[5035]: W1124 20:41:09.140637 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69d32786_a0f3_4285_be95_d3c2264cd8d1.slice/crio-72158a3990c5f807ec5619567df2b44abdf3c2a4488ed0da2f9fd4895b249329 WatchSource:0}: Error finding container 72158a3990c5f807ec5619567df2b44abdf3c2a4488ed0da2f9fd4895b249329: Status 404 returned error can't find the container with id 72158a3990c5f807ec5619567df2b44abdf3c2a4488ed0da2f9fd4895b249329 Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.945255 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92qth" event={"ID":"66eff236-3b10-4ae4-8afd-8e4257ac72de","Type":"ContainerStarted","Data":"5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438"} Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.947365 5035 generic.go:334] "Generic (PLEG): container finished" podID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerID="6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24" exitCode=0 Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.947461 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml2k" event={"ID":"5962b0dc-5936-455a-8731-12aeabbd0a72","Type":"ContainerDied","Data":"6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24"} Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.951970 5035 generic.go:334] "Generic (PLEG): container finished" podID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerID="07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f" exitCode=0 Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.952034 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lcvh9" event={"ID":"69d32786-a0f3-4285-be95-d3c2264cd8d1","Type":"ContainerDied","Data":"07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f"} Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.952084 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lcvh9" event={"ID":"69d32786-a0f3-4285-be95-d3c2264cd8d1","Type":"ContainerStarted","Data":"72158a3990c5f807ec5619567df2b44abdf3c2a4488ed0da2f9fd4895b249329"} Nov 24 20:41:09 crc kubenswrapper[5035]: I1124 20:41:09.977539 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-92qth" podStartSLOduration=3.216177812 podStartE2EDuration="6.977516716s" podCreationTimestamp="2025-11-24 20:41:03 +0000 UTC" firstStartedPulling="2025-11-24 20:41:04.856858802 +0000 UTC m=+5203.379365079" lastFinishedPulling="2025-11-24 20:41:08.618197726 +0000 UTC m=+5207.140703983" observedRunningTime="2025-11-24 20:41:09.971215817 +0000 UTC m=+5208.493722064" watchObservedRunningTime="2025-11-24 20:41:09.977516716 +0000 UTC m=+5208.500022963" Nov 24 20:41:10 crc kubenswrapper[5035]: I1124 20:41:10.963620 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lcvh9" event={"ID":"69d32786-a0f3-4285-be95-d3c2264cd8d1","Type":"ContainerStarted","Data":"e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db"} Nov 24 20:41:10 crc kubenswrapper[5035]: I1124 20:41:10.969329 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml2k" event={"ID":"5962b0dc-5936-455a-8731-12aeabbd0a72","Type":"ContainerStarted","Data":"3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9"} Nov 24 20:41:11 crc kubenswrapper[5035]: I1124 20:41:11.002283 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sml2k" podStartSLOduration=3.435587051 podStartE2EDuration="8.002261168s" podCreationTimestamp="2025-11-24 20:41:03 +0000 UTC" firstStartedPulling="2025-11-24 20:41:05.876708341 +0000 UTC m=+5204.399214598" lastFinishedPulling="2025-11-24 20:41:10.443382458 +0000 UTC m=+5208.965888715" observedRunningTime="2025-11-24 20:41:10.998536538 +0000 UTC m=+5209.521042805" watchObservedRunningTime="2025-11-24 20:41:11.002261168 +0000 UTC m=+5209.524767425" Nov 24 20:41:11 crc kubenswrapper[5035]: I1124 20:41:11.999633 5035 generic.go:334] "Generic (PLEG): container finished" podID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerID="e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db" exitCode=0 Nov 24 20:41:12 crc kubenswrapper[5035]: I1124 20:41:11.999991 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lcvh9" event={"ID":"69d32786-a0f3-4285-be95-d3c2264cd8d1","Type":"ContainerDied","Data":"e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db"} Nov 24 20:41:12 crc kubenswrapper[5035]: E1124 20:41:12.116500 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69d32786_a0f3_4285_be95_d3c2264cd8d1.slice/crio-conmon-e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db.scope\": RecentStats: unable to find data in memory cache]" Nov 24 20:41:13 crc kubenswrapper[5035]: I1124 20:41:13.012206 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lcvh9" event={"ID":"69d32786-a0f3-4285-be95-d3c2264cd8d1","Type":"ContainerStarted","Data":"7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087"} Nov 24 20:41:13 crc kubenswrapper[5035]: I1124 20:41:13.036996 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lcvh9" podStartSLOduration=2.577292403 podStartE2EDuration="5.036977514s" podCreationTimestamp="2025-11-24 20:41:08 +0000 UTC" firstStartedPulling="2025-11-24 20:41:09.953949754 +0000 UTC m=+5208.476456011" lastFinishedPulling="2025-11-24 20:41:12.413634865 +0000 UTC m=+5210.936141122" observedRunningTime="2025-11-24 20:41:13.033398078 +0000 UTC m=+5211.555904335" watchObservedRunningTime="2025-11-24 20:41:13.036977514 +0000 UTC m=+5211.559483771" Nov 24 20:41:13 crc kubenswrapper[5035]: I1124 20:41:13.744842 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:13 crc kubenswrapper[5035]: I1124 20:41:13.744902 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:13 crc kubenswrapper[5035]: I1124 20:41:13.935609 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:13 crc kubenswrapper[5035]: I1124 20:41:13.935925 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:14 crc kubenswrapper[5035]: I1124 20:41:14.796630 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-92qth" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="registry-server" probeResult="failure" output=< Nov 24 20:41:14 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:41:14 crc kubenswrapper[5035]: > Nov 24 20:41:14 crc kubenswrapper[5035]: I1124 20:41:14.979161 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-sml2k" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="registry-server" probeResult="failure" output=< Nov 24 20:41:14 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:41:14 crc kubenswrapper[5035]: > Nov 24 20:41:17 crc kubenswrapper[5035]: I1124 20:41:17.200544 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:41:17 crc kubenswrapper[5035]: E1124 20:41:17.201188 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:41:18 crc kubenswrapper[5035]: I1124 20:41:18.553639 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:18 crc kubenswrapper[5035]: I1124 20:41:18.553690 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:18 crc kubenswrapper[5035]: I1124 20:41:18.983063 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:19 crc kubenswrapper[5035]: I1124 20:41:19.110356 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:19 crc kubenswrapper[5035]: I1124 20:41:19.230821 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lcvh9"] Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.080739 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lcvh9" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="registry-server" containerID="cri-o://7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087" gracePeriod=2 Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.668327 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.766742 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-catalog-content\") pod \"69d32786-a0f3-4285-be95-d3c2264cd8d1\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.766863 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59jcf\" (UniqueName: \"kubernetes.io/projected/69d32786-a0f3-4285-be95-d3c2264cd8d1-kube-api-access-59jcf\") pod \"69d32786-a0f3-4285-be95-d3c2264cd8d1\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.766902 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-utilities\") pod \"69d32786-a0f3-4285-be95-d3c2264cd8d1\" (UID: \"69d32786-a0f3-4285-be95-d3c2264cd8d1\") " Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.767756 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-utilities" (OuterVolumeSpecName: "utilities") pod "69d32786-a0f3-4285-be95-d3c2264cd8d1" (UID: "69d32786-a0f3-4285-be95-d3c2264cd8d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.784282 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69d32786-a0f3-4285-be95-d3c2264cd8d1-kube-api-access-59jcf" (OuterVolumeSpecName: "kube-api-access-59jcf") pod "69d32786-a0f3-4285-be95-d3c2264cd8d1" (UID: "69d32786-a0f3-4285-be95-d3c2264cd8d1"). InnerVolumeSpecName "kube-api-access-59jcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.788441 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69d32786-a0f3-4285-be95-d3c2264cd8d1" (UID: "69d32786-a0f3-4285-be95-d3c2264cd8d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.869765 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.869806 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59jcf\" (UniqueName: \"kubernetes.io/projected/69d32786-a0f3-4285-be95-d3c2264cd8d1-kube-api-access-59jcf\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:21 crc kubenswrapper[5035]: I1124 20:41:21.869817 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d32786-a0f3-4285-be95-d3c2264cd8d1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.091573 5035 generic.go:334] "Generic (PLEG): container finished" podID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerID="7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087" exitCode=0 Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.091630 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lcvh9" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.091640 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lcvh9" event={"ID":"69d32786-a0f3-4285-be95-d3c2264cd8d1","Type":"ContainerDied","Data":"7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087"} Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.092037 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lcvh9" event={"ID":"69d32786-a0f3-4285-be95-d3c2264cd8d1","Type":"ContainerDied","Data":"72158a3990c5f807ec5619567df2b44abdf3c2a4488ed0da2f9fd4895b249329"} Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.092067 5035 scope.go:117] "RemoveContainer" containerID="7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.124672 5035 scope.go:117] "RemoveContainer" containerID="e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.134477 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lcvh9"] Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.152726 5035 scope.go:117] "RemoveContainer" containerID="07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.156588 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lcvh9"] Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.209191 5035 scope.go:117] "RemoveContainer" containerID="7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087" Nov 24 20:41:22 crc kubenswrapper[5035]: E1124 20:41:22.209641 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087\": container with ID starting with 7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087 not found: ID does not exist" containerID="7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.209672 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087"} err="failed to get container status \"7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087\": rpc error: code = NotFound desc = could not find container \"7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087\": container with ID starting with 7319450c2cc697dccaaf1f58f939a89addc9bb2ca4b34d583f6571b974376087 not found: ID does not exist" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.209691 5035 scope.go:117] "RemoveContainer" containerID="e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db" Nov 24 20:41:22 crc kubenswrapper[5035]: E1124 20:41:22.210094 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db\": container with ID starting with e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db not found: ID does not exist" containerID="e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.210152 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db"} err="failed to get container status \"e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db\": rpc error: code = NotFound desc = could not find container \"e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db\": container with ID starting with e9f09e35a255cd872a3792a61a8ba6edc22e44c0d852252e54ec6946f55609db not found: ID does not exist" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.210188 5035 scope.go:117] "RemoveContainer" containerID="07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f" Nov 24 20:41:22 crc kubenswrapper[5035]: E1124 20:41:22.210581 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f\": container with ID starting with 07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f not found: ID does not exist" containerID="07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.210610 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f"} err="failed to get container status \"07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f\": rpc error: code = NotFound desc = could not find container \"07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f\": container with ID starting with 07b65bf8ed9df84cd00939fdd310b5748be227bf71f18b5216d9b7f5116f216f not found: ID does not exist" Nov 24 20:41:22 crc kubenswrapper[5035]: I1124 20:41:22.214210 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" path="/var/lib/kubelet/pods/69d32786-a0f3-4285-be95-d3c2264cd8d1/volumes" Nov 24 20:41:23 crc kubenswrapper[5035]: I1124 20:41:23.792734 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:23 crc kubenswrapper[5035]: I1124 20:41:23.857842 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:23 crc kubenswrapper[5035]: I1124 20:41:23.991227 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:24 crc kubenswrapper[5035]: I1124 20:41:24.042197 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.227043 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-92qth"] Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.227607 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-92qth" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="registry-server" containerID="cri-o://5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438" gracePeriod=2 Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.804538 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.863310 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-utilities\") pod \"66eff236-3b10-4ae4-8afd-8e4257ac72de\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.863686 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhjbk\" (UniqueName: \"kubernetes.io/projected/66eff236-3b10-4ae4-8afd-8e4257ac72de-kube-api-access-zhjbk\") pod \"66eff236-3b10-4ae4-8afd-8e4257ac72de\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.863871 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-catalog-content\") pod \"66eff236-3b10-4ae4-8afd-8e4257ac72de\" (UID: \"66eff236-3b10-4ae4-8afd-8e4257ac72de\") " Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.863954 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-utilities" (OuterVolumeSpecName: "utilities") pod "66eff236-3b10-4ae4-8afd-8e4257ac72de" (UID: "66eff236-3b10-4ae4-8afd-8e4257ac72de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.864549 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.875542 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66eff236-3b10-4ae4-8afd-8e4257ac72de-kube-api-access-zhjbk" (OuterVolumeSpecName: "kube-api-access-zhjbk") pod "66eff236-3b10-4ae4-8afd-8e4257ac72de" (UID: "66eff236-3b10-4ae4-8afd-8e4257ac72de"). InnerVolumeSpecName "kube-api-access-zhjbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.914445 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66eff236-3b10-4ae4-8afd-8e4257ac72de" (UID: "66eff236-3b10-4ae4-8afd-8e4257ac72de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.966150 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhjbk\" (UniqueName: \"kubernetes.io/projected/66eff236-3b10-4ae4-8afd-8e4257ac72de-kube-api-access-zhjbk\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:25 crc kubenswrapper[5035]: I1124 20:41:25.966193 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66eff236-3b10-4ae4-8afd-8e4257ac72de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.129159 5035 generic.go:334] "Generic (PLEG): container finished" podID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerID="5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438" exitCode=0 Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.129197 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92qth" event={"ID":"66eff236-3b10-4ae4-8afd-8e4257ac72de","Type":"ContainerDied","Data":"5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438"} Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.129222 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92qth" event={"ID":"66eff236-3b10-4ae4-8afd-8e4257ac72de","Type":"ContainerDied","Data":"9eff37a834b75132e935a1c6f2a1c6b72ec8534323087a52ab3e3b8bc5b958c4"} Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.129228 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92qth" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.129239 5035 scope.go:117] "RemoveContainer" containerID="5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.161943 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-92qth"] Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.162192 5035 scope.go:117] "RemoveContainer" containerID="ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.169637 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-92qth"] Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.187219 5035 scope.go:117] "RemoveContainer" containerID="9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.212785 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" path="/var/lib/kubelet/pods/66eff236-3b10-4ae4-8afd-8e4257ac72de/volumes" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.228729 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sml2k"] Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.228971 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sml2k" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="registry-server" containerID="cri-o://3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9" gracePeriod=2 Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.262760 5035 scope.go:117] "RemoveContainer" containerID="5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438" Nov 24 20:41:26 crc kubenswrapper[5035]: E1124 20:41:26.263272 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438\": container with ID starting with 5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438 not found: ID does not exist" containerID="5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.263326 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438"} err="failed to get container status \"5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438\": rpc error: code = NotFound desc = could not find container \"5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438\": container with ID starting with 5a6be5ce1ca47aef19c5be4f8e82bba5cac5d666e4e9b2fc8248830641d85438 not found: ID does not exist" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.263365 5035 scope.go:117] "RemoveContainer" containerID="ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff" Nov 24 20:41:26 crc kubenswrapper[5035]: E1124 20:41:26.263726 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff\": container with ID starting with ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff not found: ID does not exist" containerID="ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.263757 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff"} err="failed to get container status \"ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff\": rpc error: code = NotFound desc = could not find container \"ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff\": container with ID starting with ec21de1cc9386baac2d996ac02f0aecf3e1d508bec9b6581646a2315ead66fff not found: ID does not exist" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.263781 5035 scope.go:117] "RemoveContainer" containerID="9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4" Nov 24 20:41:26 crc kubenswrapper[5035]: E1124 20:41:26.264048 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4\": container with ID starting with 9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4 not found: ID does not exist" containerID="9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.264068 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4"} err="failed to get container status \"9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4\": rpc error: code = NotFound desc = could not find container \"9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4\": container with ID starting with 9205c9fee2a7269831312d8e6d99a26a8b5ed4d687b578f515aa6f231fd92fa4 not found: ID does not exist" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.857027 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.984446 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-utilities\") pod \"5962b0dc-5936-455a-8731-12aeabbd0a72\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.984617 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tcgv\" (UniqueName: \"kubernetes.io/projected/5962b0dc-5936-455a-8731-12aeabbd0a72-kube-api-access-2tcgv\") pod \"5962b0dc-5936-455a-8731-12aeabbd0a72\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.984648 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-catalog-content\") pod \"5962b0dc-5936-455a-8731-12aeabbd0a72\" (UID: \"5962b0dc-5936-455a-8731-12aeabbd0a72\") " Nov 24 20:41:26 crc kubenswrapper[5035]: I1124 20:41:26.985730 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-utilities" (OuterVolumeSpecName: "utilities") pod "5962b0dc-5936-455a-8731-12aeabbd0a72" (UID: "5962b0dc-5936-455a-8731-12aeabbd0a72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.012112 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5962b0dc-5936-455a-8731-12aeabbd0a72-kube-api-access-2tcgv" (OuterVolumeSpecName: "kube-api-access-2tcgv") pod "5962b0dc-5936-455a-8731-12aeabbd0a72" (UID: "5962b0dc-5936-455a-8731-12aeabbd0a72"). InnerVolumeSpecName "kube-api-access-2tcgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.052443 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5962b0dc-5936-455a-8731-12aeabbd0a72" (UID: "5962b0dc-5936-455a-8731-12aeabbd0a72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.086617 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tcgv\" (UniqueName: \"kubernetes.io/projected/5962b0dc-5936-455a-8731-12aeabbd0a72-kube-api-access-2tcgv\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.086651 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.086661 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5962b0dc-5936-455a-8731-12aeabbd0a72-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.141628 5035 generic.go:334] "Generic (PLEG): container finished" podID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerID="3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9" exitCode=0 Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.141678 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sml2k" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.141676 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml2k" event={"ID":"5962b0dc-5936-455a-8731-12aeabbd0a72","Type":"ContainerDied","Data":"3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9"} Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.141824 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sml2k" event={"ID":"5962b0dc-5936-455a-8731-12aeabbd0a72","Type":"ContainerDied","Data":"7c329a53fa99f80ef6f02e30c539151babafbbb377eda08427a3ea469d529893"} Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.141872 5035 scope.go:117] "RemoveContainer" containerID="3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.172576 5035 scope.go:117] "RemoveContainer" containerID="6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.184491 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sml2k"] Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.190085 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sml2k"] Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.204962 5035 scope.go:117] "RemoveContainer" containerID="fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.247948 5035 scope.go:117] "RemoveContainer" containerID="3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9" Nov 24 20:41:27 crc kubenswrapper[5035]: E1124 20:41:27.248371 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9\": container with ID starting with 3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9 not found: ID does not exist" containerID="3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.248410 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9"} err="failed to get container status \"3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9\": rpc error: code = NotFound desc = could not find container \"3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9\": container with ID starting with 3a4e4ba343b7b1ffaaca3b521006db3039e1eff06b963db9596166da9656b6e9 not found: ID does not exist" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.248435 5035 scope.go:117] "RemoveContainer" containerID="6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24" Nov 24 20:41:27 crc kubenswrapper[5035]: E1124 20:41:27.248813 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24\": container with ID starting with 6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24 not found: ID does not exist" containerID="6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.248844 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24"} err="failed to get container status \"6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24\": rpc error: code = NotFound desc = could not find container \"6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24\": container with ID starting with 6086ce0b1c15f3653282201ca547b274e6085e7c81022edbc54c8c3d6b377a24 not found: ID does not exist" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.248867 5035 scope.go:117] "RemoveContainer" containerID="fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253" Nov 24 20:41:27 crc kubenswrapper[5035]: E1124 20:41:27.249140 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253\": container with ID starting with fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253 not found: ID does not exist" containerID="fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253" Nov 24 20:41:27 crc kubenswrapper[5035]: I1124 20:41:27.249172 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253"} err="failed to get container status \"fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253\": rpc error: code = NotFound desc = could not find container \"fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253\": container with ID starting with fbf001f4e40b78c11f36f04fefdddbe2955caddaacdaf0f6489e52c82557b253 not found: ID does not exist" Nov 24 20:41:28 crc kubenswrapper[5035]: I1124 20:41:28.213944 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" path="/var/lib/kubelet/pods/5962b0dc-5936-455a-8731-12aeabbd0a72/volumes" Nov 24 20:41:32 crc kubenswrapper[5035]: I1124 20:41:32.208571 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:41:32 crc kubenswrapper[5035]: E1124 20:41:32.209389 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:41:47 crc kubenswrapper[5035]: I1124 20:41:47.201787 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:41:47 crc kubenswrapper[5035]: E1124 20:41:47.202601 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:42:02 crc kubenswrapper[5035]: I1124 20:42:02.206711 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:42:02 crc kubenswrapper[5035]: E1124 20:42:02.207979 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:42:14 crc kubenswrapper[5035]: I1124 20:42:14.208258 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:42:14 crc kubenswrapper[5035]: E1124 20:42:14.209849 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:42:28 crc kubenswrapper[5035]: I1124 20:42:28.200795 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:42:28 crc kubenswrapper[5035]: E1124 20:42:28.201617 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:42:43 crc kubenswrapper[5035]: I1124 20:42:43.199839 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:42:43 crc kubenswrapper[5035]: E1124 20:42:43.200595 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:42:56 crc kubenswrapper[5035]: I1124 20:42:56.200532 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:42:56 crc kubenswrapper[5035]: E1124 20:42:56.201351 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.795837 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rg2th"] Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796788 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="extract-utilities" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796801 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="extract-utilities" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796826 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="extract-utilities" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796832 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="extract-utilities" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796845 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796851 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796861 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="extract-content" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796867 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="extract-content" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796879 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796884 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796893 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="extract-content" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796898 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="extract-content" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796917 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="extract-content" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796922 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="extract-content" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796937 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796944 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: E1124 20:42:59.796957 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="extract-utilities" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.796963 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="extract-utilities" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.797136 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="66eff236-3b10-4ae4-8afd-8e4257ac72de" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.797159 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5962b0dc-5936-455a-8731-12aeabbd0a72" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.797174 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="69d32786-a0f3-4285-be95-d3c2264cd8d1" containerName="registry-server" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.798803 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.802777 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rg2th"] Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.901363 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv94q\" (UniqueName: \"kubernetes.io/projected/36e43a7b-c213-457b-8129-7f64fd50ba49-kube-api-access-dv94q\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.901462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-catalog-content\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:42:59 crc kubenswrapper[5035]: I1124 20:42:59.901688 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-utilities\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.003589 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv94q\" (UniqueName: \"kubernetes.io/projected/36e43a7b-c213-457b-8129-7f64fd50ba49-kube-api-access-dv94q\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.003675 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-catalog-content\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.003727 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-utilities\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.004274 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-catalog-content\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.004416 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-utilities\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.035339 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv94q\" (UniqueName: \"kubernetes.io/projected/36e43a7b-c213-457b-8129-7f64fd50ba49-kube-api-access-dv94q\") pod \"redhat-operators-rg2th\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.131464 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:00 crc kubenswrapper[5035]: I1124 20:43:00.632821 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rg2th"] Nov 24 20:43:01 crc kubenswrapper[5035]: I1124 20:43:01.036123 5035 generic.go:334] "Generic (PLEG): container finished" podID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerID="8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af" exitCode=0 Nov 24 20:43:01 crc kubenswrapper[5035]: I1124 20:43:01.036421 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rg2th" event={"ID":"36e43a7b-c213-457b-8129-7f64fd50ba49","Type":"ContainerDied","Data":"8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af"} Nov 24 20:43:01 crc kubenswrapper[5035]: I1124 20:43:01.036496 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rg2th" event={"ID":"36e43a7b-c213-457b-8129-7f64fd50ba49","Type":"ContainerStarted","Data":"4dcfbb8b861e17ebf78cb9df3149ff2ed7c021808fe3fea61295379dbf63151c"} Nov 24 20:43:03 crc kubenswrapper[5035]: I1124 20:43:03.057068 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rg2th" event={"ID":"36e43a7b-c213-457b-8129-7f64fd50ba49","Type":"ContainerStarted","Data":"45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a"} Nov 24 20:43:08 crc kubenswrapper[5035]: I1124 20:43:08.200037 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:43:08 crc kubenswrapper[5035]: E1124 20:43:08.200876 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:43:13 crc kubenswrapper[5035]: I1124 20:43:13.203432 5035 generic.go:334] "Generic (PLEG): container finished" podID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerID="45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a" exitCode=0 Nov 24 20:43:13 crc kubenswrapper[5035]: I1124 20:43:13.203552 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rg2th" event={"ID":"36e43a7b-c213-457b-8129-7f64fd50ba49","Type":"ContainerDied","Data":"45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a"} Nov 24 20:43:15 crc kubenswrapper[5035]: I1124 20:43:15.224988 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rg2th" event={"ID":"36e43a7b-c213-457b-8129-7f64fd50ba49","Type":"ContainerStarted","Data":"227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d"} Nov 24 20:43:15 crc kubenswrapper[5035]: I1124 20:43:15.250663 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rg2th" podStartSLOduration=3.41001139 podStartE2EDuration="16.250642499s" podCreationTimestamp="2025-11-24 20:42:59 +0000 UTC" firstStartedPulling="2025-11-24 20:43:01.037525856 +0000 UTC m=+5319.560032113" lastFinishedPulling="2025-11-24 20:43:13.878156955 +0000 UTC m=+5332.400663222" observedRunningTime="2025-11-24 20:43:15.242436178 +0000 UTC m=+5333.764942435" watchObservedRunningTime="2025-11-24 20:43:15.250642499 +0000 UTC m=+5333.773148756" Nov 24 20:43:19 crc kubenswrapper[5035]: I1124 20:43:19.200614 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:43:20 crc kubenswrapper[5035]: I1124 20:43:20.132111 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:20 crc kubenswrapper[5035]: I1124 20:43:20.132764 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:20 crc kubenswrapper[5035]: I1124 20:43:20.267549 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"3276f50806c66d8b9c6e0b45a68067950ab9e5019d76beaba0eaabfe4bb15511"} Nov 24 20:43:21 crc kubenswrapper[5035]: I1124 20:43:21.205064 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rg2th" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="registry-server" probeResult="failure" output=< Nov 24 20:43:21 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 20:43:21 crc kubenswrapper[5035]: > Nov 24 20:43:30 crc kubenswrapper[5035]: I1124 20:43:30.198650 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:30 crc kubenswrapper[5035]: I1124 20:43:30.272197 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:30 crc kubenswrapper[5035]: I1124 20:43:30.993546 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rg2th"] Nov 24 20:43:31 crc kubenswrapper[5035]: I1124 20:43:31.375479 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rg2th" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="registry-server" containerID="cri-o://227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d" gracePeriod=2 Nov 24 20:43:31 crc kubenswrapper[5035]: I1124 20:43:31.951188 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.096241 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv94q\" (UniqueName: \"kubernetes.io/projected/36e43a7b-c213-457b-8129-7f64fd50ba49-kube-api-access-dv94q\") pod \"36e43a7b-c213-457b-8129-7f64fd50ba49\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.096282 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-catalog-content\") pod \"36e43a7b-c213-457b-8129-7f64fd50ba49\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.096456 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-utilities\") pod \"36e43a7b-c213-457b-8129-7f64fd50ba49\" (UID: \"36e43a7b-c213-457b-8129-7f64fd50ba49\") " Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.097389 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-utilities" (OuterVolumeSpecName: "utilities") pod "36e43a7b-c213-457b-8129-7f64fd50ba49" (UID: "36e43a7b-c213-457b-8129-7f64fd50ba49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.103503 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36e43a7b-c213-457b-8129-7f64fd50ba49-kube-api-access-dv94q" (OuterVolumeSpecName: "kube-api-access-dv94q") pod "36e43a7b-c213-457b-8129-7f64fd50ba49" (UID: "36e43a7b-c213-457b-8129-7f64fd50ba49"). InnerVolumeSpecName "kube-api-access-dv94q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.187865 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36e43a7b-c213-457b-8129-7f64fd50ba49" (UID: "36e43a7b-c213-457b-8129-7f64fd50ba49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.199102 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.199147 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36e43a7b-c213-457b-8129-7f64fd50ba49-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.199163 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv94q\" (UniqueName: \"kubernetes.io/projected/36e43a7b-c213-457b-8129-7f64fd50ba49-kube-api-access-dv94q\") on node \"crc\" DevicePath \"\"" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.386785 5035 generic.go:334] "Generic (PLEG): container finished" podID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerID="227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d" exitCode=0 Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.386862 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rg2th" event={"ID":"36e43a7b-c213-457b-8129-7f64fd50ba49","Type":"ContainerDied","Data":"227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d"} Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.386894 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rg2th" event={"ID":"36e43a7b-c213-457b-8129-7f64fd50ba49","Type":"ContainerDied","Data":"4dcfbb8b861e17ebf78cb9df3149ff2ed7c021808fe3fea61295379dbf63151c"} Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.386900 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rg2th" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.386913 5035 scope.go:117] "RemoveContainer" containerID="227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.411900 5035 scope.go:117] "RemoveContainer" containerID="45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.416945 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rg2th"] Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.424720 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rg2th"] Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.432690 5035 scope.go:117] "RemoveContainer" containerID="8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.490384 5035 scope.go:117] "RemoveContainer" containerID="227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d" Nov 24 20:43:32 crc kubenswrapper[5035]: E1124 20:43:32.490992 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d\": container with ID starting with 227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d not found: ID does not exist" containerID="227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.491027 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d"} err="failed to get container status \"227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d\": rpc error: code = NotFound desc = could not find container \"227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d\": container with ID starting with 227ce789e7fc1dadf1ca3ce269e4981509f76a71fcbdff534cde3512f815de3d not found: ID does not exist" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.491048 5035 scope.go:117] "RemoveContainer" containerID="45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a" Nov 24 20:43:32 crc kubenswrapper[5035]: E1124 20:43:32.491376 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a\": container with ID starting with 45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a not found: ID does not exist" containerID="45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.491401 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a"} err="failed to get container status \"45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a\": rpc error: code = NotFound desc = could not find container \"45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a\": container with ID starting with 45bb214b43e4c6fc6061fb6a6da4d85f39b697fd4b56807ba15e774f68c9776a not found: ID does not exist" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.491415 5035 scope.go:117] "RemoveContainer" containerID="8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af" Nov 24 20:43:32 crc kubenswrapper[5035]: E1124 20:43:32.491726 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af\": container with ID starting with 8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af not found: ID does not exist" containerID="8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af" Nov 24 20:43:32 crc kubenswrapper[5035]: I1124 20:43:32.491747 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af"} err="failed to get container status \"8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af\": rpc error: code = NotFound desc = could not find container \"8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af\": container with ID starting with 8aa3e0ab552b7c69f7d85142db3aada1184c5f5524e4539cf812d963e4f0c0af not found: ID does not exist" Nov 24 20:43:34 crc kubenswrapper[5035]: I1124 20:43:34.214554 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" path="/var/lib/kubelet/pods/36e43a7b-c213-457b-8129-7f64fd50ba49/volumes" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.148538 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq"] Nov 24 20:45:00 crc kubenswrapper[5035]: E1124 20:45:00.149560 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="registry-server" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.149578 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="registry-server" Nov 24 20:45:00 crc kubenswrapper[5035]: E1124 20:45:00.149596 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="extract-utilities" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.149604 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="extract-utilities" Nov 24 20:45:00 crc kubenswrapper[5035]: E1124 20:45:00.149631 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="extract-content" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.149638 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="extract-content" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.149884 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e43a7b-c213-457b-8129-7f64fd50ba49" containerName="registry-server" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.150691 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.152533 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.153074 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.160013 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq"] Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.251463 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86203792-eae4-48cf-999d-11b281aa8d72-secret-volume\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.251822 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86203792-eae4-48cf-999d-11b281aa8d72-config-volume\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.252034 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vccm\" (UniqueName: \"kubernetes.io/projected/86203792-eae4-48cf-999d-11b281aa8d72-kube-api-access-5vccm\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.354093 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vccm\" (UniqueName: \"kubernetes.io/projected/86203792-eae4-48cf-999d-11b281aa8d72-kube-api-access-5vccm\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.354228 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86203792-eae4-48cf-999d-11b281aa8d72-secret-volume\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.354255 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86203792-eae4-48cf-999d-11b281aa8d72-config-volume\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.355170 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86203792-eae4-48cf-999d-11b281aa8d72-config-volume\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.360998 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86203792-eae4-48cf-999d-11b281aa8d72-secret-volume\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.374451 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vccm\" (UniqueName: \"kubernetes.io/projected/86203792-eae4-48cf-999d-11b281aa8d72-kube-api-access-5vccm\") pod \"collect-profiles-29400285-zcznq\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.484167 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:00 crc kubenswrapper[5035]: I1124 20:45:00.973537 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq"] Nov 24 20:45:01 crc kubenswrapper[5035]: I1124 20:45:01.221639 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" event={"ID":"86203792-eae4-48cf-999d-11b281aa8d72","Type":"ContainerStarted","Data":"b75c048a0754b3542afae516d0940ade1f0996423711c1e707a729946fe4e221"} Nov 24 20:45:01 crc kubenswrapper[5035]: I1124 20:45:01.221990 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" event={"ID":"86203792-eae4-48cf-999d-11b281aa8d72","Type":"ContainerStarted","Data":"dd137983bd1a27a3edb2a465221a01cccba5aeefa5f6cce28c5dfd3c491bf2eb"} Nov 24 20:45:01 crc kubenswrapper[5035]: I1124 20:45:01.245818 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" podStartSLOduration=1.245766319 podStartE2EDuration="1.245766319s" podCreationTimestamp="2025-11-24 20:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:45:01.24242048 +0000 UTC m=+5439.764926757" watchObservedRunningTime="2025-11-24 20:45:01.245766319 +0000 UTC m=+5439.768272586" Nov 24 20:45:02 crc kubenswrapper[5035]: I1124 20:45:02.230895 5035 generic.go:334] "Generic (PLEG): container finished" podID="86203792-eae4-48cf-999d-11b281aa8d72" containerID="b75c048a0754b3542afae516d0940ade1f0996423711c1e707a729946fe4e221" exitCode=0 Nov 24 20:45:02 crc kubenswrapper[5035]: I1124 20:45:02.231219 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" event={"ID":"86203792-eae4-48cf-999d-11b281aa8d72","Type":"ContainerDied","Data":"b75c048a0754b3542afae516d0940ade1f0996423711c1e707a729946fe4e221"} Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.656136 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.743819 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86203792-eae4-48cf-999d-11b281aa8d72-secret-volume\") pod \"86203792-eae4-48cf-999d-11b281aa8d72\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.744192 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vccm\" (UniqueName: \"kubernetes.io/projected/86203792-eae4-48cf-999d-11b281aa8d72-kube-api-access-5vccm\") pod \"86203792-eae4-48cf-999d-11b281aa8d72\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.744511 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86203792-eae4-48cf-999d-11b281aa8d72-config-volume\") pod \"86203792-eae4-48cf-999d-11b281aa8d72\" (UID: \"86203792-eae4-48cf-999d-11b281aa8d72\") " Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.745538 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86203792-eae4-48cf-999d-11b281aa8d72-config-volume" (OuterVolumeSpecName: "config-volume") pod "86203792-eae4-48cf-999d-11b281aa8d72" (UID: "86203792-eae4-48cf-999d-11b281aa8d72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.753188 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86203792-eae4-48cf-999d-11b281aa8d72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "86203792-eae4-48cf-999d-11b281aa8d72" (UID: "86203792-eae4-48cf-999d-11b281aa8d72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.753380 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86203792-eae4-48cf-999d-11b281aa8d72-kube-api-access-5vccm" (OuterVolumeSpecName: "kube-api-access-5vccm") pod "86203792-eae4-48cf-999d-11b281aa8d72" (UID: "86203792-eae4-48cf-999d-11b281aa8d72"). InnerVolumeSpecName "kube-api-access-5vccm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.846897 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vccm\" (UniqueName: \"kubernetes.io/projected/86203792-eae4-48cf-999d-11b281aa8d72-kube-api-access-5vccm\") on node \"crc\" DevicePath \"\"" Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.846928 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86203792-eae4-48cf-999d-11b281aa8d72-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:45:03 crc kubenswrapper[5035]: I1124 20:45:03.846936 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86203792-eae4-48cf-999d-11b281aa8d72-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 20:45:04 crc kubenswrapper[5035]: I1124 20:45:04.252540 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" event={"ID":"86203792-eae4-48cf-999d-11b281aa8d72","Type":"ContainerDied","Data":"dd137983bd1a27a3edb2a465221a01cccba5aeefa5f6cce28c5dfd3c491bf2eb"} Nov 24 20:45:04 crc kubenswrapper[5035]: I1124 20:45:04.252691 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd137983bd1a27a3edb2a465221a01cccba5aeefa5f6cce28c5dfd3c491bf2eb" Nov 24 20:45:04 crc kubenswrapper[5035]: I1124 20:45:04.252850 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400285-zcznq" Nov 24 20:45:04 crc kubenswrapper[5035]: I1124 20:45:04.725857 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv"] Nov 24 20:45:04 crc kubenswrapper[5035]: I1124 20:45:04.739193 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400240-tprvv"] Nov 24 20:45:06 crc kubenswrapper[5035]: I1124 20:45:06.211317 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72106bf6-5bff-4412-b58e-47c869df5338" path="/var/lib/kubelet/pods/72106bf6-5bff-4412-b58e-47c869df5338/volumes" Nov 24 20:45:32 crc kubenswrapper[5035]: I1124 20:45:32.126830 5035 scope.go:117] "RemoveContainer" containerID="c4e87ae1e98c69667633be750652a0ce9517239f1e37682e94a4679d23ef1f40" Nov 24 20:45:45 crc kubenswrapper[5035]: I1124 20:45:45.236981 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:45:45 crc kubenswrapper[5035]: I1124 20:45:45.237547 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:46:15 crc kubenswrapper[5035]: I1124 20:46:15.234204 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:46:15 crc kubenswrapper[5035]: I1124 20:46:15.234765 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:46:45 crc kubenswrapper[5035]: I1124 20:46:45.233848 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:46:45 crc kubenswrapper[5035]: I1124 20:46:45.236208 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:46:45 crc kubenswrapper[5035]: I1124 20:46:45.236779 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:46:45 crc kubenswrapper[5035]: I1124 20:46:45.238970 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3276f50806c66d8b9c6e0b45a68067950ab9e5019d76beaba0eaabfe4bb15511"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:46:45 crc kubenswrapper[5035]: I1124 20:46:45.239644 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://3276f50806c66d8b9c6e0b45a68067950ab9e5019d76beaba0eaabfe4bb15511" gracePeriod=600 Nov 24 20:46:46 crc kubenswrapper[5035]: I1124 20:46:46.306142 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="3276f50806c66d8b9c6e0b45a68067950ab9e5019d76beaba0eaabfe4bb15511" exitCode=0 Nov 24 20:46:46 crc kubenswrapper[5035]: I1124 20:46:46.306203 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"3276f50806c66d8b9c6e0b45a68067950ab9e5019d76beaba0eaabfe4bb15511"} Nov 24 20:46:46 crc kubenswrapper[5035]: I1124 20:46:46.307081 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6"} Nov 24 20:46:46 crc kubenswrapper[5035]: I1124 20:46:46.307131 5035 scope.go:117] "RemoveContainer" containerID="d3c842ab7d4653fccf56f5baa845f93fbd670aa547c49ea61a45c41bbcf9268a" Nov 24 20:48:45 crc kubenswrapper[5035]: I1124 20:48:45.234161 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:48:45 crc kubenswrapper[5035]: I1124 20:48:45.234876 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:49:15 crc kubenswrapper[5035]: I1124 20:49:15.234662 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:49:15 crc kubenswrapper[5035]: I1124 20:49:15.235274 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:49:45 crc kubenswrapper[5035]: I1124 20:49:45.234964 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:49:45 crc kubenswrapper[5035]: I1124 20:49:45.235834 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:49:45 crc kubenswrapper[5035]: I1124 20:49:45.235902 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:49:45 crc kubenswrapper[5035]: I1124 20:49:45.237062 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:49:45 crc kubenswrapper[5035]: I1124 20:49:45.237170 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" gracePeriod=600 Nov 24 20:49:45 crc kubenswrapper[5035]: E1124 20:49:45.371657 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:49:45 crc kubenswrapper[5035]: E1124 20:49:45.472734 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11fbb8c9_66d2_4fdd_bb91_b4dfb6ea9361.slice/crio-fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11fbb8c9_66d2_4fdd_bb91_b4dfb6ea9361.slice/crio-conmon-fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6.scope\": RecentStats: unable to find data in memory cache]" Nov 24 20:49:46 crc kubenswrapper[5035]: I1124 20:49:46.154626 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" exitCode=0 Nov 24 20:49:46 crc kubenswrapper[5035]: I1124 20:49:46.154653 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6"} Nov 24 20:49:46 crc kubenswrapper[5035]: I1124 20:49:46.154926 5035 scope.go:117] "RemoveContainer" containerID="3276f50806c66d8b9c6e0b45a68067950ab9e5019d76beaba0eaabfe4bb15511" Nov 24 20:49:46 crc kubenswrapper[5035]: I1124 20:49:46.155641 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:49:46 crc kubenswrapper[5035]: E1124 20:49:46.155922 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:50:00 crc kubenswrapper[5035]: I1124 20:50:00.201406 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:50:00 crc kubenswrapper[5035]: E1124 20:50:00.202696 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:50:12 crc kubenswrapper[5035]: I1124 20:50:12.205487 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:50:12 crc kubenswrapper[5035]: E1124 20:50:12.206265 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:50:23 crc kubenswrapper[5035]: I1124 20:50:23.201334 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:50:23 crc kubenswrapper[5035]: E1124 20:50:23.202142 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:50:35 crc kubenswrapper[5035]: I1124 20:50:35.199747 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:50:35 crc kubenswrapper[5035]: E1124 20:50:35.200521 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:50:46 crc kubenswrapper[5035]: I1124 20:50:46.202467 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:50:46 crc kubenswrapper[5035]: E1124 20:50:46.203263 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:51:00 crc kubenswrapper[5035]: I1124 20:51:00.200908 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:51:00 crc kubenswrapper[5035]: E1124 20:51:00.201833 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:51:11 crc kubenswrapper[5035]: I1124 20:51:11.200796 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:51:11 crc kubenswrapper[5035]: E1124 20:51:11.201733 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.143973 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-72xtv"] Nov 24 20:51:20 crc kubenswrapper[5035]: E1124 20:51:20.145316 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86203792-eae4-48cf-999d-11b281aa8d72" containerName="collect-profiles" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.145342 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="86203792-eae4-48cf-999d-11b281aa8d72" containerName="collect-profiles" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.145678 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="86203792-eae4-48cf-999d-11b281aa8d72" containerName="collect-profiles" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.148047 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.156850 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-72xtv"] Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.285884 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-catalog-content\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.286073 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwrz6\" (UniqueName: \"kubernetes.io/projected/f326ddb7-417c-402f-9ddb-b4e36d41ad25-kube-api-access-gwrz6\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.286143 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-utilities\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.387806 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwrz6\" (UniqueName: \"kubernetes.io/projected/f326ddb7-417c-402f-9ddb-b4e36d41ad25-kube-api-access-gwrz6\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.387928 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-utilities\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.388097 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-catalog-content\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.388534 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-utilities\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.388933 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-catalog-content\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.420110 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwrz6\" (UniqueName: \"kubernetes.io/projected/f326ddb7-417c-402f-9ddb-b4e36d41ad25-kube-api-access-gwrz6\") pod \"certified-operators-72xtv\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:20 crc kubenswrapper[5035]: I1124 20:51:20.476675 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:21 crc kubenswrapper[5035]: I1124 20:51:21.099447 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-72xtv"] Nov 24 20:51:21 crc kubenswrapper[5035]: I1124 20:51:21.536646 5035 generic.go:334] "Generic (PLEG): container finished" podID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerID="1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6" exitCode=0 Nov 24 20:51:21 crc kubenswrapper[5035]: I1124 20:51:21.536710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72xtv" event={"ID":"f326ddb7-417c-402f-9ddb-b4e36d41ad25","Type":"ContainerDied","Data":"1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6"} Nov 24 20:51:21 crc kubenswrapper[5035]: I1124 20:51:21.537072 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72xtv" event={"ID":"f326ddb7-417c-402f-9ddb-b4e36d41ad25","Type":"ContainerStarted","Data":"84ba249297a2d5153964e7be045467a71378c2de4080feaa2bc4e0ad3c17264a"} Nov 24 20:51:21 crc kubenswrapper[5035]: I1124 20:51:21.538422 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:51:22 crc kubenswrapper[5035]: I1124 20:51:22.217785 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:51:22 crc kubenswrapper[5035]: E1124 20:51:22.219063 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:51:23 crc kubenswrapper[5035]: I1124 20:51:23.556472 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72xtv" event={"ID":"f326ddb7-417c-402f-9ddb-b4e36d41ad25","Type":"ContainerStarted","Data":"752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418"} Nov 24 20:51:26 crc kubenswrapper[5035]: I1124 20:51:26.589656 5035 generic.go:334] "Generic (PLEG): container finished" podID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerID="752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418" exitCode=0 Nov 24 20:51:26 crc kubenswrapper[5035]: I1124 20:51:26.590434 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72xtv" event={"ID":"f326ddb7-417c-402f-9ddb-b4e36d41ad25","Type":"ContainerDied","Data":"752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418"} Nov 24 20:51:27 crc kubenswrapper[5035]: I1124 20:51:27.603554 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72xtv" event={"ID":"f326ddb7-417c-402f-9ddb-b4e36d41ad25","Type":"ContainerStarted","Data":"f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc"} Nov 24 20:51:27 crc kubenswrapper[5035]: I1124 20:51:27.650472 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-72xtv" podStartSLOduration=2.089991979 podStartE2EDuration="7.650451819s" podCreationTimestamp="2025-11-24 20:51:20 +0000 UTC" firstStartedPulling="2025-11-24 20:51:21.538022274 +0000 UTC m=+5820.060528531" lastFinishedPulling="2025-11-24 20:51:27.098482114 +0000 UTC m=+5825.620988371" observedRunningTime="2025-11-24 20:51:27.640459281 +0000 UTC m=+5826.162965538" watchObservedRunningTime="2025-11-24 20:51:27.650451819 +0000 UTC m=+5826.172958086" Nov 24 20:51:30 crc kubenswrapper[5035]: I1124 20:51:30.479103 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:30 crc kubenswrapper[5035]: I1124 20:51:30.479561 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:30 crc kubenswrapper[5035]: I1124 20:51:30.529580 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:34 crc kubenswrapper[5035]: I1124 20:51:34.201096 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:51:34 crc kubenswrapper[5035]: E1124 20:51:34.202119 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:51:40 crc kubenswrapper[5035]: I1124 20:51:40.555372 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:40 crc kubenswrapper[5035]: I1124 20:51:40.621587 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-72xtv"] Nov 24 20:51:40 crc kubenswrapper[5035]: I1124 20:51:40.739446 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-72xtv" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="registry-server" containerID="cri-o://f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc" gracePeriod=2 Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.274634 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.461785 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-utilities\") pod \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.461855 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-catalog-content\") pod \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.461897 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwrz6\" (UniqueName: \"kubernetes.io/projected/f326ddb7-417c-402f-9ddb-b4e36d41ad25-kube-api-access-gwrz6\") pod \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\" (UID: \"f326ddb7-417c-402f-9ddb-b4e36d41ad25\") " Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.463274 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-utilities" (OuterVolumeSpecName: "utilities") pod "f326ddb7-417c-402f-9ddb-b4e36d41ad25" (UID: "f326ddb7-417c-402f-9ddb-b4e36d41ad25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.474504 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f326ddb7-417c-402f-9ddb-b4e36d41ad25-kube-api-access-gwrz6" (OuterVolumeSpecName: "kube-api-access-gwrz6") pod "f326ddb7-417c-402f-9ddb-b4e36d41ad25" (UID: "f326ddb7-417c-402f-9ddb-b4e36d41ad25"). InnerVolumeSpecName "kube-api-access-gwrz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.509939 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f326ddb7-417c-402f-9ddb-b4e36d41ad25" (UID: "f326ddb7-417c-402f-9ddb-b4e36d41ad25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.564011 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.564943 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f326ddb7-417c-402f-9ddb-b4e36d41ad25-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.565028 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwrz6\" (UniqueName: \"kubernetes.io/projected/f326ddb7-417c-402f-9ddb-b4e36d41ad25-kube-api-access-gwrz6\") on node \"crc\" DevicePath \"\"" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.749137 5035 generic.go:334] "Generic (PLEG): container finished" podID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerID="f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc" exitCode=0 Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.749355 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72xtv" event={"ID":"f326ddb7-417c-402f-9ddb-b4e36d41ad25","Type":"ContainerDied","Data":"f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc"} Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.749572 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-72xtv" event={"ID":"f326ddb7-417c-402f-9ddb-b4e36d41ad25","Type":"ContainerDied","Data":"84ba249297a2d5153964e7be045467a71378c2de4080feaa2bc4e0ad3c17264a"} Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.749657 5035 scope.go:117] "RemoveContainer" containerID="f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.749449 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-72xtv" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.788615 5035 scope.go:117] "RemoveContainer" containerID="752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.799564 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-72xtv"] Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.811161 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-72xtv"] Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.831827 5035 scope.go:117] "RemoveContainer" containerID="1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.889700 5035 scope.go:117] "RemoveContainer" containerID="f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc" Nov 24 20:51:41 crc kubenswrapper[5035]: E1124 20:51:41.891215 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc\": container with ID starting with f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc not found: ID does not exist" containerID="f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.891269 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc"} err="failed to get container status \"f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc\": rpc error: code = NotFound desc = could not find container \"f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc\": container with ID starting with f3721f3aa9a2974e9bccfa3ff9612592eff4e3ac4ac0bf5ba002b1281e617fbc not found: ID does not exist" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.891325 5035 scope.go:117] "RemoveContainer" containerID="752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418" Nov 24 20:51:41 crc kubenswrapper[5035]: E1124 20:51:41.891769 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418\": container with ID starting with 752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418 not found: ID does not exist" containerID="752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.891876 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418"} err="failed to get container status \"752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418\": rpc error: code = NotFound desc = could not find container \"752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418\": container with ID starting with 752144dbc64a742d0fd0f49fb0c633fe71998b7bf27f347c2864c5cd0680b418 not found: ID does not exist" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.891958 5035 scope.go:117] "RemoveContainer" containerID="1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6" Nov 24 20:51:41 crc kubenswrapper[5035]: E1124 20:51:41.892287 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6\": container with ID starting with 1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6 not found: ID does not exist" containerID="1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6" Nov 24 20:51:41 crc kubenswrapper[5035]: I1124 20:51:41.892438 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6"} err="failed to get container status \"1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6\": rpc error: code = NotFound desc = could not find container \"1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6\": container with ID starting with 1867e168aad847732f5b5a7a13ddb1b5862e5ae5f043c028771840a0980c66d6 not found: ID does not exist" Nov 24 20:51:42 crc kubenswrapper[5035]: I1124 20:51:42.213033 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" path="/var/lib/kubelet/pods/f326ddb7-417c-402f-9ddb-b4e36d41ad25/volumes" Nov 24 20:51:49 crc kubenswrapper[5035]: I1124 20:51:49.200010 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:51:49 crc kubenswrapper[5035]: E1124 20:51:49.201008 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:52:01 crc kubenswrapper[5035]: I1124 20:52:01.200597 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:52:01 crc kubenswrapper[5035]: E1124 20:52:01.201580 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:52:14 crc kubenswrapper[5035]: I1124 20:52:14.200019 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:52:14 crc kubenswrapper[5035]: E1124 20:52:14.200666 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.201105 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:52:25 crc kubenswrapper[5035]: E1124 20:52:25.202249 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.938719 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-29lbc"] Nov 24 20:52:25 crc kubenswrapper[5035]: E1124 20:52:25.939136 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="extract-content" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.939153 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="extract-content" Nov 24 20:52:25 crc kubenswrapper[5035]: E1124 20:52:25.939181 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="extract-utilities" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.939187 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="extract-utilities" Nov 24 20:52:25 crc kubenswrapper[5035]: E1124 20:52:25.939201 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="registry-server" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.939207 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="registry-server" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.939399 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f326ddb7-417c-402f-9ddb-b4e36d41ad25" containerName="registry-server" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.940684 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.954437 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29lbc"] Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.997703 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-utilities\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.997773 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-catalog-content\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:25 crc kubenswrapper[5035]: I1124 20:52:25.998105 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2j78\" (UniqueName: \"kubernetes.io/projected/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-kube-api-access-m2j78\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.099938 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-utilities\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.099988 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-catalog-content\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.100084 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2j78\" (UniqueName: \"kubernetes.io/projected/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-kube-api-access-m2j78\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.100767 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-utilities\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.100937 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-catalog-content\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.127130 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2j78\" (UniqueName: \"kubernetes.io/projected/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-kube-api-access-m2j78\") pod \"redhat-marketplace-29lbc\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.269205 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:26 crc kubenswrapper[5035]: I1124 20:52:26.760445 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29lbc"] Nov 24 20:52:27 crc kubenswrapper[5035]: I1124 20:52:27.250487 5035 generic.go:334] "Generic (PLEG): container finished" podID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerID="7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d" exitCode=0 Nov 24 20:52:27 crc kubenswrapper[5035]: I1124 20:52:27.250593 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29lbc" event={"ID":"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1","Type":"ContainerDied","Data":"7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d"} Nov 24 20:52:27 crc kubenswrapper[5035]: I1124 20:52:27.250733 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29lbc" event={"ID":"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1","Type":"ContainerStarted","Data":"694c04016bf29e836cd8ef8c0d0862f2cf133f14224aaaf312c6d90df5f3c2f7"} Nov 24 20:52:28 crc kubenswrapper[5035]: I1124 20:52:28.263385 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29lbc" event={"ID":"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1","Type":"ContainerStarted","Data":"0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1"} Nov 24 20:52:29 crc kubenswrapper[5035]: I1124 20:52:29.276284 5035 generic.go:334] "Generic (PLEG): container finished" podID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerID="0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1" exitCode=0 Nov 24 20:52:29 crc kubenswrapper[5035]: I1124 20:52:29.278333 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29lbc" event={"ID":"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1","Type":"ContainerDied","Data":"0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1"} Nov 24 20:52:30 crc kubenswrapper[5035]: I1124 20:52:30.289124 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29lbc" event={"ID":"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1","Type":"ContainerStarted","Data":"c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f"} Nov 24 20:52:30 crc kubenswrapper[5035]: I1124 20:52:30.318606 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-29lbc" podStartSLOduration=2.754258557 podStartE2EDuration="5.31858993s" podCreationTimestamp="2025-11-24 20:52:25 +0000 UTC" firstStartedPulling="2025-11-24 20:52:27.252724773 +0000 UTC m=+5885.775231020" lastFinishedPulling="2025-11-24 20:52:29.817056136 +0000 UTC m=+5888.339562393" observedRunningTime="2025-11-24 20:52:30.312820454 +0000 UTC m=+5888.835326741" watchObservedRunningTime="2025-11-24 20:52:30.31858993 +0000 UTC m=+5888.841096187" Nov 24 20:52:36 crc kubenswrapper[5035]: I1124 20:52:36.270613 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:36 crc kubenswrapper[5035]: I1124 20:52:36.271187 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:36 crc kubenswrapper[5035]: I1124 20:52:36.348415 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:36 crc kubenswrapper[5035]: I1124 20:52:36.412615 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:36 crc kubenswrapper[5035]: I1124 20:52:36.590009 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29lbc"] Nov 24 20:52:38 crc kubenswrapper[5035]: I1124 20:52:38.370775 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-29lbc" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="registry-server" containerID="cri-o://c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f" gracePeriod=2 Nov 24 20:52:38 crc kubenswrapper[5035]: I1124 20:52:38.885908 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.030777 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-catalog-content\") pod \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.030879 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-utilities\") pod \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.031107 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2j78\" (UniqueName: \"kubernetes.io/projected/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-kube-api-access-m2j78\") pod \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\" (UID: \"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1\") " Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.032421 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-utilities" (OuterVolumeSpecName: "utilities") pod "ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" (UID: "ea5321b4-d5c8-4d05-96f1-eb165b2f95d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.037656 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-kube-api-access-m2j78" (OuterVolumeSpecName: "kube-api-access-m2j78") pod "ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" (UID: "ea5321b4-d5c8-4d05-96f1-eb165b2f95d1"). InnerVolumeSpecName "kube-api-access-m2j78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.055487 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" (UID: "ea5321b4-d5c8-4d05-96f1-eb165b2f95d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.134480 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.134549 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.134573 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2j78\" (UniqueName: \"kubernetes.io/projected/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1-kube-api-access-m2j78\") on node \"crc\" DevicePath \"\"" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.201665 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:52:39 crc kubenswrapper[5035]: E1124 20:52:39.201996 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.380962 5035 generic.go:334] "Generic (PLEG): container finished" podID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerID="c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f" exitCode=0 Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.381016 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29lbc" event={"ID":"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1","Type":"ContainerDied","Data":"c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f"} Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.381041 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29lbc" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.381047 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29lbc" event={"ID":"ea5321b4-d5c8-4d05-96f1-eb165b2f95d1","Type":"ContainerDied","Data":"694c04016bf29e836cd8ef8c0d0862f2cf133f14224aaaf312c6d90df5f3c2f7"} Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.381084 5035 scope.go:117] "RemoveContainer" containerID="c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.424869 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29lbc"] Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.426835 5035 scope.go:117] "RemoveContainer" containerID="0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.434170 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-29lbc"] Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.453335 5035 scope.go:117] "RemoveContainer" containerID="7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.507414 5035 scope.go:117] "RemoveContainer" containerID="c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f" Nov 24 20:52:39 crc kubenswrapper[5035]: E1124 20:52:39.507953 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f\": container with ID starting with c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f not found: ID does not exist" containerID="c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.507993 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f"} err="failed to get container status \"c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f\": rpc error: code = NotFound desc = could not find container \"c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f\": container with ID starting with c999621a0bf98c3b72007d11c4796e6afc8722f5963d8ca5b2b0eaff0d06653f not found: ID does not exist" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.508020 5035 scope.go:117] "RemoveContainer" containerID="0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1" Nov 24 20:52:39 crc kubenswrapper[5035]: E1124 20:52:39.508486 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1\": container with ID starting with 0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1 not found: ID does not exist" containerID="0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.508548 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1"} err="failed to get container status \"0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1\": rpc error: code = NotFound desc = could not find container \"0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1\": container with ID starting with 0f90e1d47c4b090cf4d24863edadd70b70d1ef0f5b644de56761160b759925d1 not found: ID does not exist" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.508586 5035 scope.go:117] "RemoveContainer" containerID="7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d" Nov 24 20:52:39 crc kubenswrapper[5035]: E1124 20:52:39.509041 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d\": container with ID starting with 7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d not found: ID does not exist" containerID="7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d" Nov 24 20:52:39 crc kubenswrapper[5035]: I1124 20:52:39.509065 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d"} err="failed to get container status \"7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d\": rpc error: code = NotFound desc = could not find container \"7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d\": container with ID starting with 7a77afdf4e393aca3dc4f4c4a427a13d9f378ab262e55a82cf67b0253724fa1d not found: ID does not exist" Nov 24 20:52:40 crc kubenswrapper[5035]: I1124 20:52:40.220506 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" path="/var/lib/kubelet/pods/ea5321b4-d5c8-4d05-96f1-eb165b2f95d1/volumes" Nov 24 20:52:52 crc kubenswrapper[5035]: I1124 20:52:52.226544 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:52:52 crc kubenswrapper[5035]: E1124 20:52:52.228597 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:53:03 crc kubenswrapper[5035]: I1124 20:53:03.200908 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:53:03 crc kubenswrapper[5035]: E1124 20:53:03.202000 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:53:14 crc kubenswrapper[5035]: I1124 20:53:14.201228 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:53:14 crc kubenswrapper[5035]: E1124 20:53:14.202153 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:53:25 crc kubenswrapper[5035]: I1124 20:53:25.200911 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:53:25 crc kubenswrapper[5035]: E1124 20:53:25.201542 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:53:36 crc kubenswrapper[5035]: I1124 20:53:36.200459 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:53:36 crc kubenswrapper[5035]: E1124 20:53:36.201375 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:53:48 crc kubenswrapper[5035]: I1124 20:53:48.201147 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:53:48 crc kubenswrapper[5035]: E1124 20:53:48.202353 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:53:59 crc kubenswrapper[5035]: I1124 20:53:59.200180 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:53:59 crc kubenswrapper[5035]: E1124 20:53:59.201382 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:54:13 crc kubenswrapper[5035]: I1124 20:54:13.200710 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:54:13 crc kubenswrapper[5035]: E1124 20:54:13.202103 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.539190 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xgsbs"] Nov 24 20:54:14 crc kubenswrapper[5035]: E1124 20:54:14.540327 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="extract-utilities" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.540353 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="extract-utilities" Nov 24 20:54:14 crc kubenswrapper[5035]: E1124 20:54:14.540385 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="registry-server" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.540398 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="registry-server" Nov 24 20:54:14 crc kubenswrapper[5035]: E1124 20:54:14.540429 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="extract-content" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.540445 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="extract-content" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.540852 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea5321b4-d5c8-4d05-96f1-eb165b2f95d1" containerName="registry-server" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.543702 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.555512 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xgsbs"] Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.717609 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qdl2\" (UniqueName: \"kubernetes.io/projected/68e334b0-890b-4002-95fd-94811284fbcc-kube-api-access-7qdl2\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.717772 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-utilities\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.717953 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-catalog-content\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.819579 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qdl2\" (UniqueName: \"kubernetes.io/projected/68e334b0-890b-4002-95fd-94811284fbcc-kube-api-access-7qdl2\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.819825 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-utilities\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.820469 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-utilities\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.820681 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-catalog-content\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.821105 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-catalog-content\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.848928 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qdl2\" (UniqueName: \"kubernetes.io/projected/68e334b0-890b-4002-95fd-94811284fbcc-kube-api-access-7qdl2\") pod \"redhat-operators-xgsbs\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:14 crc kubenswrapper[5035]: I1124 20:54:14.869988 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:15 crc kubenswrapper[5035]: I1124 20:54:15.311329 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xgsbs"] Nov 24 20:54:15 crc kubenswrapper[5035]: I1124 20:54:15.425484 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgsbs" event={"ID":"68e334b0-890b-4002-95fd-94811284fbcc","Type":"ContainerStarted","Data":"3bb875b4c31730b76554066fa00d522a17fe5e86bd0fd98325e41d51fcaea63e"} Nov 24 20:54:16 crc kubenswrapper[5035]: I1124 20:54:16.434385 5035 generic.go:334] "Generic (PLEG): container finished" podID="68e334b0-890b-4002-95fd-94811284fbcc" containerID="fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373" exitCode=0 Nov 24 20:54:16 crc kubenswrapper[5035]: I1124 20:54:16.434648 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgsbs" event={"ID":"68e334b0-890b-4002-95fd-94811284fbcc","Type":"ContainerDied","Data":"fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373"} Nov 24 20:54:18 crc kubenswrapper[5035]: I1124 20:54:18.457682 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgsbs" event={"ID":"68e334b0-890b-4002-95fd-94811284fbcc","Type":"ContainerStarted","Data":"76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe"} Nov 24 20:54:24 crc kubenswrapper[5035]: I1124 20:54:24.201641 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:54:24 crc kubenswrapper[5035]: E1124 20:54:24.203047 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:54:25 crc kubenswrapper[5035]: I1124 20:54:25.527808 5035 generic.go:334] "Generic (PLEG): container finished" podID="68e334b0-890b-4002-95fd-94811284fbcc" containerID="76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe" exitCode=0 Nov 24 20:54:25 crc kubenswrapper[5035]: I1124 20:54:25.527915 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgsbs" event={"ID":"68e334b0-890b-4002-95fd-94811284fbcc","Type":"ContainerDied","Data":"76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe"} Nov 24 20:54:27 crc kubenswrapper[5035]: I1124 20:54:27.550717 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgsbs" event={"ID":"68e334b0-890b-4002-95fd-94811284fbcc","Type":"ContainerStarted","Data":"4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e"} Nov 24 20:54:27 crc kubenswrapper[5035]: I1124 20:54:27.575378 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xgsbs" podStartSLOduration=3.647253132 podStartE2EDuration="13.575364772s" podCreationTimestamp="2025-11-24 20:54:14 +0000 UTC" firstStartedPulling="2025-11-24 20:54:16.4365357 +0000 UTC m=+5994.959041957" lastFinishedPulling="2025-11-24 20:54:26.36464734 +0000 UTC m=+6004.887153597" observedRunningTime="2025-11-24 20:54:27.567829799 +0000 UTC m=+6006.090336066" watchObservedRunningTime="2025-11-24 20:54:27.575364772 +0000 UTC m=+6006.097871019" Nov 24 20:54:34 crc kubenswrapper[5035]: I1124 20:54:34.871365 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:34 crc kubenswrapper[5035]: I1124 20:54:34.871923 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:34 crc kubenswrapper[5035]: I1124 20:54:34.923312 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:35 crc kubenswrapper[5035]: I1124 20:54:35.200098 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:54:35 crc kubenswrapper[5035]: E1124 20:54:35.200448 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 20:54:35 crc kubenswrapper[5035]: I1124 20:54:35.695307 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:35 crc kubenswrapper[5035]: I1124 20:54:35.750070 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xgsbs"] Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.065087 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xgsbs" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="registry-server" containerID="cri-o://4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e" gracePeriod=2 Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.691175 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.856590 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-utilities\") pod \"68e334b0-890b-4002-95fd-94811284fbcc\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.856647 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qdl2\" (UniqueName: \"kubernetes.io/projected/68e334b0-890b-4002-95fd-94811284fbcc-kube-api-access-7qdl2\") pod \"68e334b0-890b-4002-95fd-94811284fbcc\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.856768 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-catalog-content\") pod \"68e334b0-890b-4002-95fd-94811284fbcc\" (UID: \"68e334b0-890b-4002-95fd-94811284fbcc\") " Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.857965 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-utilities" (OuterVolumeSpecName: "utilities") pod "68e334b0-890b-4002-95fd-94811284fbcc" (UID: "68e334b0-890b-4002-95fd-94811284fbcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.861697 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68e334b0-890b-4002-95fd-94811284fbcc-kube-api-access-7qdl2" (OuterVolumeSpecName: "kube-api-access-7qdl2") pod "68e334b0-890b-4002-95fd-94811284fbcc" (UID: "68e334b0-890b-4002-95fd-94811284fbcc"). InnerVolumeSpecName "kube-api-access-7qdl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.948527 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68e334b0-890b-4002-95fd-94811284fbcc" (UID: "68e334b0-890b-4002-95fd-94811284fbcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.958838 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.958872 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qdl2\" (UniqueName: \"kubernetes.io/projected/68e334b0-890b-4002-95fd-94811284fbcc-kube-api-access-7qdl2\") on node \"crc\" DevicePath \"\"" Nov 24 20:54:38 crc kubenswrapper[5035]: I1124 20:54:38.958885 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68e334b0-890b-4002-95fd-94811284fbcc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.079214 5035 generic.go:334] "Generic (PLEG): container finished" podID="68e334b0-890b-4002-95fd-94811284fbcc" containerID="4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e" exitCode=0 Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.079260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgsbs" event={"ID":"68e334b0-890b-4002-95fd-94811284fbcc","Type":"ContainerDied","Data":"4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e"} Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.079278 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgsbs" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.079339 5035 scope.go:117] "RemoveContainer" containerID="4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.079325 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgsbs" event={"ID":"68e334b0-890b-4002-95fd-94811284fbcc","Type":"ContainerDied","Data":"3bb875b4c31730b76554066fa00d522a17fe5e86bd0fd98325e41d51fcaea63e"} Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.111765 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xgsbs"] Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.115843 5035 scope.go:117] "RemoveContainer" containerID="76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.123475 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xgsbs"] Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.142372 5035 scope.go:117] "RemoveContainer" containerID="fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.196257 5035 scope.go:117] "RemoveContainer" containerID="4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e" Nov 24 20:54:39 crc kubenswrapper[5035]: E1124 20:54:39.196933 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e\": container with ID starting with 4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e not found: ID does not exist" containerID="4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.196978 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e"} err="failed to get container status \"4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e\": rpc error: code = NotFound desc = could not find container \"4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e\": container with ID starting with 4341ef81772a2fcfc746508359f8910afd028b1a7a88dc00f935b45583507a7e not found: ID does not exist" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.197014 5035 scope.go:117] "RemoveContainer" containerID="76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe" Nov 24 20:54:39 crc kubenswrapper[5035]: E1124 20:54:39.197563 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe\": container with ID starting with 76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe not found: ID does not exist" containerID="76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.197621 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe"} err="failed to get container status \"76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe\": rpc error: code = NotFound desc = could not find container \"76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe\": container with ID starting with 76df67c7ac7d249c724578dd505415c6403fe36798ca1014a8a407886490b5fe not found: ID does not exist" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.197656 5035 scope.go:117] "RemoveContainer" containerID="fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373" Nov 24 20:54:39 crc kubenswrapper[5035]: E1124 20:54:39.197926 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373\": container with ID starting with fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373 not found: ID does not exist" containerID="fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373" Nov 24 20:54:39 crc kubenswrapper[5035]: I1124 20:54:39.197951 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373"} err="failed to get container status \"fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373\": rpc error: code = NotFound desc = could not find container \"fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373\": container with ID starting with fb97a7271574fe6cafffa882d3490de7b2073c8fd100d62daaf8d413da71d373 not found: ID does not exist" Nov 24 20:54:40 crc kubenswrapper[5035]: I1124 20:54:40.217525 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68e334b0-890b-4002-95fd-94811284fbcc" path="/var/lib/kubelet/pods/68e334b0-890b-4002-95fd-94811284fbcc/volumes" Nov 24 20:54:46 crc kubenswrapper[5035]: I1124 20:54:46.201080 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:54:47 crc kubenswrapper[5035]: I1124 20:54:47.208407 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"ec02eac9e2bb0061a2c002a2e8f9ba07191c078cfe0a44c8ad6165a85a41d8a2"} Nov 24 20:55:44 crc kubenswrapper[5035]: I1124 20:55:44.838373 5035 generic.go:334] "Generic (PLEG): container finished" podID="03d6dde6-e977-4ec1-a23d-1642e6c23029" containerID="06c50d2f475343fb268f78f8b509ef1a07f252e6f5c5d715bf3d531053913a30" exitCode=1 Nov 24 20:55:44 crc kubenswrapper[5035]: I1124 20:55:44.838465 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"03d6dde6-e977-4ec1-a23d-1642e6c23029","Type":"ContainerDied","Data":"06c50d2f475343fb268f78f8b509ef1a07f252e6f5c5d715bf3d531053913a30"} Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.347390 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.438407 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Nov 24 20:55:46 crc kubenswrapper[5035]: E1124 20:55:46.438775 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="registry-server" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.438790 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="registry-server" Nov 24 20:55:46 crc kubenswrapper[5035]: E1124 20:55:46.438813 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d6dde6-e977-4ec1-a23d-1642e6c23029" containerName="tempest-tests-tempest-tests-runner" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.438821 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d6dde6-e977-4ec1-a23d-1642e6c23029" containerName="tempest-tests-tempest-tests-runner" Nov 24 20:55:46 crc kubenswrapper[5035]: E1124 20:55:46.438837 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="extract-content" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.438843 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="extract-content" Nov 24 20:55:46 crc kubenswrapper[5035]: E1124 20:55:46.438854 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="extract-utilities" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.438860 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="extract-utilities" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.439024 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="68e334b0-890b-4002-95fd-94811284fbcc" containerName="registry-server" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.439039 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d6dde6-e977-4ec1-a23d-1642e6c23029" containerName="tempest-tests-tempest-tests-runner" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.439826 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.441762 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.442014 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453375 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ceph\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453458 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6rxt\" (UniqueName: \"kubernetes.io/projected/03d6dde6-e977-4ec1-a23d-1642e6c23029-kube-api-access-j6rxt\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453496 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config-secret\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453556 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-config-data\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453601 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ca-certs\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453637 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453717 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453765 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ssh-key\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453937 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-temporary\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.453964 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-workdir\") pod \"03d6dde6-e977-4ec1-a23d-1642e6c23029\" (UID: \"03d6dde6-e977-4ec1-a23d-1642e6c23029\") " Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.454626 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-config-data" (OuterVolumeSpecName: "config-data") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.455383 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.459525 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.460883 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.465889 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.466659 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d6dde6-e977-4ec1-a23d-1642e6c23029-kube-api-access-j6rxt" (OuterVolumeSpecName: "kube-api-access-j6rxt") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "kube-api-access-j6rxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.468119 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ceph" (OuterVolumeSpecName: "ceph") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.510916 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.511447 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.523078 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.526884 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "03d6dde6-e977-4ec1-a23d-1642e6c23029" (UID: "03d6dde6-e977-4ec1-a23d-1642e6c23029"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.557713 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.557805 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.557856 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.557912 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj96d\" (UniqueName: \"kubernetes.io/projected/c6b4d79c-6eea-4a14-bea9-5074054b09ca-kube-api-access-xj96d\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.557957 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.557987 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558038 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558064 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558101 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558187 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558326 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558354 5035 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558370 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558387 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558404 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558423 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/03d6dde6-e977-4ec1-a23d-1642e6c23029-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558440 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558453 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6rxt\" (UniqueName: \"kubernetes.io/projected/03d6dde6-e977-4ec1-a23d-1642e6c23029-kube-api-access-j6rxt\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.558465 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/03d6dde6-e977-4ec1-a23d-1642e6c23029-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.587355 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660371 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660427 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660464 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj96d\" (UniqueName: \"kubernetes.io/projected/c6b4d79c-6eea-4a14-bea9-5074054b09ca-kube-api-access-xj96d\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660497 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660533 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660552 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660568 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660592 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.660668 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.661363 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.661774 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.661785 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.662486 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.666338 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.666633 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.667119 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.667181 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.682997 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj96d\" (UniqueName: \"kubernetes.io/projected/c6b4d79c-6eea-4a14-bea9-5074054b09ca-kube-api-access-xj96d\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.781055 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.869269 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"03d6dde6-e977-4ec1-a23d-1642e6c23029","Type":"ContainerDied","Data":"25d4449e6036685fad57b15761165d9327cea44573d1ab2ce44a350f68963d09"} Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.869386 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25d4449e6036685fad57b15761165d9327cea44573d1ab2ce44a350f68963d09" Nov 24 20:55:46 crc kubenswrapper[5035]: I1124 20:55:46.869492 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Nov 24 20:55:47 crc kubenswrapper[5035]: I1124 20:55:47.213795 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Nov 24 20:55:47 crc kubenswrapper[5035]: W1124 20:55:47.219602 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6b4d79c_6eea_4a14_bea9_5074054b09ca.slice/crio-972ea67c825a2c47afcfd861a075e0726a101cc2f221b7d78b611a6ab5018d48 WatchSource:0}: Error finding container 972ea67c825a2c47afcfd861a075e0726a101cc2f221b7d78b611a6ab5018d48: Status 404 returned error can't find the container with id 972ea67c825a2c47afcfd861a075e0726a101cc2f221b7d78b611a6ab5018d48 Nov 24 20:55:47 crc kubenswrapper[5035]: I1124 20:55:47.878933 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"c6b4d79c-6eea-4a14-bea9-5074054b09ca","Type":"ContainerStarted","Data":"972ea67c825a2c47afcfd861a075e0726a101cc2f221b7d78b611a6ab5018d48"} Nov 24 20:55:48 crc kubenswrapper[5035]: I1124 20:55:48.894624 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"c6b4d79c-6eea-4a14-bea9-5074054b09ca","Type":"ContainerStarted","Data":"eace29e7ceff61465ced9f8a70941217c347c68813160b4c24c54d65b32d28aa"} Nov 24 20:55:48 crc kubenswrapper[5035]: I1124 20:55:48.927774 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-test" podStartSLOduration=2.927744146 podStartE2EDuration="2.927744146s" podCreationTimestamp="2025-11-24 20:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 20:55:48.925861965 +0000 UTC m=+6087.448368282" watchObservedRunningTime="2025-11-24 20:55:48.927744146 +0000 UTC m=+6087.450250443" Nov 24 20:57:15 crc kubenswrapper[5035]: I1124 20:57:15.234159 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:57:15 crc kubenswrapper[5035]: I1124 20:57:15.235083 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:57:45 crc kubenswrapper[5035]: I1124 20:57:45.234555 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:57:45 crc kubenswrapper[5035]: I1124 20:57:45.235129 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.267278 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-96s7h"] Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.271737 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.300679 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-96s7h"] Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.365932 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-utilities\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.365996 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-catalog-content\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.366253 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m85jk\" (UniqueName: \"kubernetes.io/projected/7fc79ae2-0c2f-479d-b859-1f32594c1674-kube-api-access-m85jk\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.467877 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-utilities\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.467938 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-catalog-content\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.468126 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m85jk\" (UniqueName: \"kubernetes.io/projected/7fc79ae2-0c2f-479d-b859-1f32594c1674-kube-api-access-m85jk\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.469531 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-catalog-content\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.472088 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-utilities\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.498866 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m85jk\" (UniqueName: \"kubernetes.io/projected/7fc79ae2-0c2f-479d-b859-1f32594c1674-kube-api-access-m85jk\") pod \"community-operators-96s7h\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:54 crc kubenswrapper[5035]: I1124 20:57:54.600743 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:57:55 crc kubenswrapper[5035]: W1124 20:57:55.189895 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fc79ae2_0c2f_479d_b859_1f32594c1674.slice/crio-cdb12eb79c4a728028a3b73722c19bc07f5a97fa1a836ede29290f9cc9bf75cd WatchSource:0}: Error finding container cdb12eb79c4a728028a3b73722c19bc07f5a97fa1a836ede29290f9cc9bf75cd: Status 404 returned error can't find the container with id cdb12eb79c4a728028a3b73722c19bc07f5a97fa1a836ede29290f9cc9bf75cd Nov 24 20:57:55 crc kubenswrapper[5035]: I1124 20:57:55.191229 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-96s7h"] Nov 24 20:57:55 crc kubenswrapper[5035]: I1124 20:57:55.354502 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-96s7h" event={"ID":"7fc79ae2-0c2f-479d-b859-1f32594c1674","Type":"ContainerStarted","Data":"cdb12eb79c4a728028a3b73722c19bc07f5a97fa1a836ede29290f9cc9bf75cd"} Nov 24 20:57:56 crc kubenswrapper[5035]: I1124 20:57:56.369462 5035 generic.go:334] "Generic (PLEG): container finished" podID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerID="0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2" exitCode=0 Nov 24 20:57:56 crc kubenswrapper[5035]: I1124 20:57:56.369597 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-96s7h" event={"ID":"7fc79ae2-0c2f-479d-b859-1f32594c1674","Type":"ContainerDied","Data":"0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2"} Nov 24 20:57:56 crc kubenswrapper[5035]: I1124 20:57:56.375965 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 20:57:57 crc kubenswrapper[5035]: I1124 20:57:57.384507 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-96s7h" event={"ID":"7fc79ae2-0c2f-479d-b859-1f32594c1674","Type":"ContainerStarted","Data":"666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66"} Nov 24 20:57:58 crc kubenswrapper[5035]: E1124 20:57:58.637952 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fc79ae2_0c2f_479d_b859_1f32594c1674.slice/crio-666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fc79ae2_0c2f_479d_b859_1f32594c1674.slice/crio-conmon-666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66.scope\": RecentStats: unable to find data in memory cache]" Nov 24 20:57:59 crc kubenswrapper[5035]: I1124 20:57:59.416468 5035 generic.go:334] "Generic (PLEG): container finished" podID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerID="666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66" exitCode=0 Nov 24 20:57:59 crc kubenswrapper[5035]: I1124 20:57:59.416550 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-96s7h" event={"ID":"7fc79ae2-0c2f-479d-b859-1f32594c1674","Type":"ContainerDied","Data":"666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66"} Nov 24 20:58:00 crc kubenswrapper[5035]: I1124 20:58:00.431077 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-96s7h" event={"ID":"7fc79ae2-0c2f-479d-b859-1f32594c1674","Type":"ContainerStarted","Data":"185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19"} Nov 24 20:58:00 crc kubenswrapper[5035]: I1124 20:58:00.450835 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-96s7h" podStartSLOduration=3.013331495 podStartE2EDuration="6.450817673s" podCreationTimestamp="2025-11-24 20:57:54 +0000 UTC" firstStartedPulling="2025-11-24 20:57:56.375567131 +0000 UTC m=+6214.898073428" lastFinishedPulling="2025-11-24 20:57:59.813053339 +0000 UTC m=+6218.335559606" observedRunningTime="2025-11-24 20:58:00.446875657 +0000 UTC m=+6218.969381924" watchObservedRunningTime="2025-11-24 20:58:00.450817673 +0000 UTC m=+6218.973323940" Nov 24 20:58:04 crc kubenswrapper[5035]: I1124 20:58:04.602411 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:58:04 crc kubenswrapper[5035]: I1124 20:58:04.603138 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:58:04 crc kubenswrapper[5035]: I1124 20:58:04.700793 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:58:05 crc kubenswrapper[5035]: I1124 20:58:05.570385 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:58:05 crc kubenswrapper[5035]: I1124 20:58:05.652325 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-96s7h"] Nov 24 20:58:07 crc kubenswrapper[5035]: I1124 20:58:07.506128 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-96s7h" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="registry-server" containerID="cri-o://185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19" gracePeriod=2 Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.069327 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.169501 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-utilities\") pod \"7fc79ae2-0c2f-479d-b859-1f32594c1674\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.170642 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-utilities" (OuterVolumeSpecName: "utilities") pod "7fc79ae2-0c2f-479d-b859-1f32594c1674" (UID: "7fc79ae2-0c2f-479d-b859-1f32594c1674"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.170775 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-catalog-content\") pod \"7fc79ae2-0c2f-479d-b859-1f32594c1674\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.170838 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m85jk\" (UniqueName: \"kubernetes.io/projected/7fc79ae2-0c2f-479d-b859-1f32594c1674-kube-api-access-m85jk\") pod \"7fc79ae2-0c2f-479d-b859-1f32594c1674\" (UID: \"7fc79ae2-0c2f-479d-b859-1f32594c1674\") " Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.172270 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.179244 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc79ae2-0c2f-479d-b859-1f32594c1674-kube-api-access-m85jk" (OuterVolumeSpecName: "kube-api-access-m85jk") pod "7fc79ae2-0c2f-479d-b859-1f32594c1674" (UID: "7fc79ae2-0c2f-479d-b859-1f32594c1674"). InnerVolumeSpecName "kube-api-access-m85jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.275141 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m85jk\" (UniqueName: \"kubernetes.io/projected/7fc79ae2-0c2f-479d-b859-1f32594c1674-kube-api-access-m85jk\") on node \"crc\" DevicePath \"\"" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.356262 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fc79ae2-0c2f-479d-b859-1f32594c1674" (UID: "7fc79ae2-0c2f-479d-b859-1f32594c1674"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.376945 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc79ae2-0c2f-479d-b859-1f32594c1674-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.521680 5035 generic.go:334] "Generic (PLEG): container finished" podID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerID="185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19" exitCode=0 Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.521760 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-96s7h" event={"ID":"7fc79ae2-0c2f-479d-b859-1f32594c1674","Type":"ContainerDied","Data":"185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19"} Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.521801 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-96s7h" event={"ID":"7fc79ae2-0c2f-479d-b859-1f32594c1674","Type":"ContainerDied","Data":"cdb12eb79c4a728028a3b73722c19bc07f5a97fa1a836ede29290f9cc9bf75cd"} Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.521832 5035 scope.go:117] "RemoveContainer" containerID="185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.522021 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-96s7h" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.553798 5035 scope.go:117] "RemoveContainer" containerID="666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.587323 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-96s7h"] Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.588690 5035 scope.go:117] "RemoveContainer" containerID="0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.601792 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-96s7h"] Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.666581 5035 scope.go:117] "RemoveContainer" containerID="185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19" Nov 24 20:58:08 crc kubenswrapper[5035]: E1124 20:58:08.667522 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19\": container with ID starting with 185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19 not found: ID does not exist" containerID="185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.667587 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19"} err="failed to get container status \"185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19\": rpc error: code = NotFound desc = could not find container \"185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19\": container with ID starting with 185d193bd9759ab962f817bead224bc1c178daf3adfabff321cf9612c24a6e19 not found: ID does not exist" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.667617 5035 scope.go:117] "RemoveContainer" containerID="666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66" Nov 24 20:58:08 crc kubenswrapper[5035]: E1124 20:58:08.668330 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66\": container with ID starting with 666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66 not found: ID does not exist" containerID="666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.668368 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66"} err="failed to get container status \"666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66\": rpc error: code = NotFound desc = could not find container \"666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66\": container with ID starting with 666deb332319e903ada851d8837c2e84ae3918b78ccb3ccf9f64189b3ba6eb66 not found: ID does not exist" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.668396 5035 scope.go:117] "RemoveContainer" containerID="0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2" Nov 24 20:58:08 crc kubenswrapper[5035]: E1124 20:58:08.668870 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2\": container with ID starting with 0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2 not found: ID does not exist" containerID="0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2" Nov 24 20:58:08 crc kubenswrapper[5035]: I1124 20:58:08.668912 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2"} err="failed to get container status \"0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2\": rpc error: code = NotFound desc = could not find container \"0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2\": container with ID starting with 0845858f5652c206a4b0ad2549cf4e2f76ca15c8b08ecd512c7879ec88d62ba2 not found: ID does not exist" Nov 24 20:58:10 crc kubenswrapper[5035]: I1124 20:58:10.211229 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" path="/var/lib/kubelet/pods/7fc79ae2-0c2f-479d-b859-1f32594c1674/volumes" Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.234631 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.234950 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.234995 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.235729 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec02eac9e2bb0061a2c002a2e8f9ba07191c078cfe0a44c8ad6165a85a41d8a2"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.235820 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://ec02eac9e2bb0061a2c002a2e8f9ba07191c078cfe0a44c8ad6165a85a41d8a2" gracePeriod=600 Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.599690 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="ec02eac9e2bb0061a2c002a2e8f9ba07191c078cfe0a44c8ad6165a85a41d8a2" exitCode=0 Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.599732 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"ec02eac9e2bb0061a2c002a2e8f9ba07191c078cfe0a44c8ad6165a85a41d8a2"} Nov 24 20:58:15 crc kubenswrapper[5035]: I1124 20:58:15.600055 5035 scope.go:117] "RemoveContainer" containerID="fecbe39de1a252abdf02dc80ab22683df0e8d95d7b2bfacf957072c12654b7b6" Nov 24 20:58:16 crc kubenswrapper[5035]: I1124 20:58:16.612149 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3"} Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.172747 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg"] Nov 24 21:00:00 crc kubenswrapper[5035]: E1124 21:00:00.173903 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="registry-server" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.173926 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="registry-server" Nov 24 21:00:00 crc kubenswrapper[5035]: E1124 21:00:00.173946 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="extract-utilities" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.173960 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="extract-utilities" Nov 24 21:00:00 crc kubenswrapper[5035]: E1124 21:00:00.173993 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="extract-content" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.174005 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="extract-content" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.174392 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc79ae2-0c2f-479d-b859-1f32594c1674" containerName="registry-server" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.175374 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.179446 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.179518 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.186063 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg"] Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.239458 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4zr4\" (UniqueName: \"kubernetes.io/projected/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-kube-api-access-c4zr4\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.239875 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-config-volume\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.239997 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-secret-volume\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.344089 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4zr4\" (UniqueName: \"kubernetes.io/projected/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-kube-api-access-c4zr4\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.344407 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-config-volume\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.344562 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-secret-volume\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.347342 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-config-volume\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.360388 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-secret-volume\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.368998 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4zr4\" (UniqueName: \"kubernetes.io/projected/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-kube-api-access-c4zr4\") pod \"collect-profiles-29400300-9frvg\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.506790 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:00 crc kubenswrapper[5035]: I1124 21:00:00.805742 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg"] Nov 24 21:00:01 crc kubenswrapper[5035]: I1124 21:00:01.814271 5035 generic.go:334] "Generic (PLEG): container finished" podID="4ec7dded-5fb2-4f9e-986c-c5b1d054e03b" containerID="b418138bea7fe36b156665a1c52d17d7c5e7d918beb1fc0d256dc15fe1537f88" exitCode=0 Nov 24 21:00:01 crc kubenswrapper[5035]: I1124 21:00:01.814568 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" event={"ID":"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b","Type":"ContainerDied","Data":"b418138bea7fe36b156665a1c52d17d7c5e7d918beb1fc0d256dc15fe1537f88"} Nov 24 21:00:01 crc kubenswrapper[5035]: I1124 21:00:01.814832 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" event={"ID":"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b","Type":"ContainerStarted","Data":"75f7ce8c009806a4cab9a432d2361071c465bf46fad613bbfa4ee6a8b4351805"} Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.162254 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.200111 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-secret-volume\") pod \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.200166 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-config-volume\") pod \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.200287 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4zr4\" (UniqueName: \"kubernetes.io/projected/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-kube-api-access-c4zr4\") pod \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\" (UID: \"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b\") " Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.201936 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-config-volume" (OuterVolumeSpecName: "config-volume") pod "4ec7dded-5fb2-4f9e-986c-c5b1d054e03b" (UID: "4ec7dded-5fb2-4f9e-986c-c5b1d054e03b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.207176 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4ec7dded-5fb2-4f9e-986c-c5b1d054e03b" (UID: "4ec7dded-5fb2-4f9e-986c-c5b1d054e03b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.207406 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-kube-api-access-c4zr4" (OuterVolumeSpecName: "kube-api-access-c4zr4") pod "4ec7dded-5fb2-4f9e-986c-c5b1d054e03b" (UID: "4ec7dded-5fb2-4f9e-986c-c5b1d054e03b"). InnerVolumeSpecName "kube-api-access-c4zr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.302623 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.302652 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.302663 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4zr4\" (UniqueName: \"kubernetes.io/projected/4ec7dded-5fb2-4f9e-986c-c5b1d054e03b-kube-api-access-c4zr4\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.831665 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" event={"ID":"4ec7dded-5fb2-4f9e-986c-c5b1d054e03b","Type":"ContainerDied","Data":"75f7ce8c009806a4cab9a432d2361071c465bf46fad613bbfa4ee6a8b4351805"} Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.832001 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75f7ce8c009806a4cab9a432d2361071c465bf46fad613bbfa4ee6a8b4351805" Nov 24 21:00:03 crc kubenswrapper[5035]: I1124 21:00:03.831794 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400300-9frvg" Nov 24 21:00:04 crc kubenswrapper[5035]: I1124 21:00:04.269469 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj"] Nov 24 21:00:04 crc kubenswrapper[5035]: I1124 21:00:04.275912 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400255-vwsmj"] Nov 24 21:00:06 crc kubenswrapper[5035]: I1124 21:00:06.211534 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ae512c2-b562-4020-822f-c3e97c4caf95" path="/var/lib/kubelet/pods/8ae512c2-b562-4020-822f-c3e97c4caf95/volumes" Nov 24 21:00:15 crc kubenswrapper[5035]: I1124 21:00:15.234078 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:00:15 crc kubenswrapper[5035]: I1124 21:00:15.234964 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:00:18 crc kubenswrapper[5035]: I1124 21:00:18.992775 5035 generic.go:334] "Generic (PLEG): container finished" podID="c6b4d79c-6eea-4a14-bea9-5074054b09ca" containerID="eace29e7ceff61465ced9f8a70941217c347c68813160b4c24c54d65b32d28aa" exitCode=0 Nov 24 21:00:18 crc kubenswrapper[5035]: I1124 21:00:18.992922 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"c6b4d79c-6eea-4a14-bea9-5074054b09ca","Type":"ContainerDied","Data":"eace29e7ceff61465ced9f8a70941217c347c68813160b4c24c54d65b32d28aa"} Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.500998 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.696691 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ssh-key\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.696813 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-temporary\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.696856 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config-secret\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.696939 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ceph\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.696964 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-workdir\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.696996 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ca-certs\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.697017 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.697051 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj96d\" (UniqueName: \"kubernetes.io/projected/c6b4d79c-6eea-4a14-bea9-5074054b09ca-kube-api-access-xj96d\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.697176 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.697274 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-config-data\") pod \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\" (UID: \"c6b4d79c-6eea-4a14-bea9-5074054b09ca\") " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.697964 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-config-data" (OuterVolumeSpecName: "config-data") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.698049 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.698571 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.698837 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.712697 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.712812 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b4d79c-6eea-4a14-bea9-5074054b09ca-kube-api-access-xj96d" (OuterVolumeSpecName: "kube-api-access-xj96d") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "kube-api-access-xj96d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.715580 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ceph" (OuterVolumeSpecName: "ceph") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.728869 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.782301 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.790445 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.800429 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.806536 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.806572 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c6b4d79c-6eea-4a14-bea9-5074054b09ca-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.806583 5035 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.806592 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj96d\" (UniqueName: \"kubernetes.io/projected/c6b4d79c-6eea-4a14-bea9-5074054b09ca-kube-api-access-xj96d\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.806615 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.806623 5035 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.806632 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.817055 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c6b4d79c-6eea-4a14-bea9-5074054b09ca" (UID: "c6b4d79c-6eea-4a14-bea9-5074054b09ca"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.848637 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.908514 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c6b4d79c-6eea-4a14-bea9-5074054b09ca-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:20 crc kubenswrapper[5035]: I1124 21:00:20.908561 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 24 21:00:21 crc kubenswrapper[5035]: I1124 21:00:21.016677 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"c6b4d79c-6eea-4a14-bea9-5074054b09ca","Type":"ContainerDied","Data":"972ea67c825a2c47afcfd861a075e0726a101cc2f221b7d78b611a6ab5018d48"} Nov 24 21:00:21 crc kubenswrapper[5035]: I1124 21:00:21.016727 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="972ea67c825a2c47afcfd861a075e0726a101cc2f221b7d78b611a6ab5018d48" Nov 24 21:00:21 crc kubenswrapper[5035]: I1124 21:00:21.016773 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.445383 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 21:00:28 crc kubenswrapper[5035]: E1124 21:00:28.447111 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec7dded-5fb2-4f9e-986c-c5b1d054e03b" containerName="collect-profiles" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.447133 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec7dded-5fb2-4f9e-986c-c5b1d054e03b" containerName="collect-profiles" Nov 24 21:00:28 crc kubenswrapper[5035]: E1124 21:00:28.447161 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b4d79c-6eea-4a14-bea9-5074054b09ca" containerName="tempest-tests-tempest-tests-runner" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.447170 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b4d79c-6eea-4a14-bea9-5074054b09ca" containerName="tempest-tests-tempest-tests-runner" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.447499 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6b4d79c-6eea-4a14-bea9-5074054b09ca" containerName="tempest-tests-tempest-tests-runner" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.447528 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec7dded-5fb2-4f9e-986c-c5b1d054e03b" containerName="collect-profiles" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.448659 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.456474 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bz72s" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.465784 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.576847 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b504ae82-14cc-45d4-8eac-9c49b8ab7172\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.576920 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ck58\" (UniqueName: \"kubernetes.io/projected/b504ae82-14cc-45d4-8eac-9c49b8ab7172-kube-api-access-7ck58\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b504ae82-14cc-45d4-8eac-9c49b8ab7172\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.678925 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b504ae82-14cc-45d4-8eac-9c49b8ab7172\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.679008 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ck58\" (UniqueName: \"kubernetes.io/projected/b504ae82-14cc-45d4-8eac-9c49b8ab7172-kube-api-access-7ck58\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b504ae82-14cc-45d4-8eac-9c49b8ab7172\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.679937 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b504ae82-14cc-45d4-8eac-9c49b8ab7172\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.705389 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ck58\" (UniqueName: \"kubernetes.io/projected/b504ae82-14cc-45d4-8eac-9c49b8ab7172-kube-api-access-7ck58\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b504ae82-14cc-45d4-8eac-9c49b8ab7172\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.740460 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b504ae82-14cc-45d4-8eac-9c49b8ab7172\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:28 crc kubenswrapper[5035]: I1124 21:00:28.771675 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 21:00:29 crc kubenswrapper[5035]: I1124 21:00:29.252242 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 21:00:29 crc kubenswrapper[5035]: W1124 21:00:29.255532 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb504ae82_14cc_45d4_8eac_9c49b8ab7172.slice/crio-750d39ad1e4ab8ad5c218782b1ab34d4c1764ff6dab41ad803d0c0c2be0312df WatchSource:0}: Error finding container 750d39ad1e4ab8ad5c218782b1ab34d4c1764ff6dab41ad803d0c0c2be0312df: Status 404 returned error can't find the container with id 750d39ad1e4ab8ad5c218782b1ab34d4c1764ff6dab41ad803d0c0c2be0312df Nov 24 21:00:30 crc kubenswrapper[5035]: I1124 21:00:30.127696 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b504ae82-14cc-45d4-8eac-9c49b8ab7172","Type":"ContainerStarted","Data":"750d39ad1e4ab8ad5c218782b1ab34d4c1764ff6dab41ad803d0c0c2be0312df"} Nov 24 21:00:32 crc kubenswrapper[5035]: I1124 21:00:32.155104 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b504ae82-14cc-45d4-8eac-9c49b8ab7172","Type":"ContainerStarted","Data":"c28842204a29eb777b0a674746883e88718fd77746ba65d567e16064f52ebc21"} Nov 24 21:00:32 crc kubenswrapper[5035]: I1124 21:00:32.181909 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.152016898 podStartE2EDuration="4.181881255s" podCreationTimestamp="2025-11-24 21:00:28 +0000 UTC" firstStartedPulling="2025-11-24 21:00:29.260502802 +0000 UTC m=+6367.783009059" lastFinishedPulling="2025-11-24 21:00:31.290367149 +0000 UTC m=+6369.812873416" observedRunningTime="2025-11-24 21:00:32.169956136 +0000 UTC m=+6370.692462473" watchObservedRunningTime="2025-11-24 21:00:32.181881255 +0000 UTC m=+6370.704387562" Nov 24 21:00:32 crc kubenswrapper[5035]: I1124 21:00:32.989826 5035 scope.go:117] "RemoveContainer" containerID="371bb4464f503221e5dff224c0edd7a25a8e8534facc83920940756c55dba90c" Nov 24 21:00:45 crc kubenswrapper[5035]: I1124 21:00:45.234905 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:00:45 crc kubenswrapper[5035]: I1124 21:00:45.235945 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.366981 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.370029 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.375027 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.375687 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-public-key" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.375952 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-config" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.376202 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-private-key" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.376551 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"tobiko-secret" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.402738 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.416113 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.416258 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.416561 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.519236 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smkwb\" (UniqueName: \"kubernetes.io/projected/3b0c51fa-ab78-4018-a361-f02d53b92345-kube-api-access-smkwb\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.519431 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.519483 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.519628 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520077 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520340 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520423 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520659 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520743 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520839 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520892 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.520998 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.521779 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.522110 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.531444 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623230 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623476 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623518 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623603 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623650 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623687 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623740 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623831 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smkwb\" (UniqueName: \"kubernetes.io/projected/3b0c51fa-ab78-4018-a361-f02d53b92345-kube-api-access-smkwb\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.623921 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.624151 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.624606 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.625438 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.626553 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.626764 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.630052 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.630329 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.631950 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.660933 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smkwb\" (UniqueName: \"kubernetes.io/projected/3b0c51fa-ab78-4018-a361-f02d53b92345-kube-api-access-smkwb\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.667674 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:47 crc kubenswrapper[5035]: I1124 21:00:47.714598 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:00:48 crc kubenswrapper[5035]: I1124 21:00:48.327185 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Nov 24 21:00:48 crc kubenswrapper[5035]: W1124 21:00:48.330321 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b0c51fa_ab78_4018_a361_f02d53b92345.slice/crio-648f05ec6b3c8cfc57b50c1454d8e6fb55927443fa2032d9c96d710a5e489ed2 WatchSource:0}: Error finding container 648f05ec6b3c8cfc57b50c1454d8e6fb55927443fa2032d9c96d710a5e489ed2: Status 404 returned error can't find the container with id 648f05ec6b3c8cfc57b50c1454d8e6fb55927443fa2032d9c96d710a5e489ed2 Nov 24 21:00:48 crc kubenswrapper[5035]: I1124 21:00:48.343421 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"3b0c51fa-ab78-4018-a361-f02d53b92345","Type":"ContainerStarted","Data":"648f05ec6b3c8cfc57b50c1454d8e6fb55927443fa2032d9c96d710a5e489ed2"} Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.149859 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29400301-dxpzw"] Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.151924 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.160250 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400301-dxpzw"] Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.212048 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-combined-ca-bundle\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.212123 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw8wn\" (UniqueName: \"kubernetes.io/projected/46fd0927-55d9-4806-b7ad-2d3633aad62a-kube-api-access-xw8wn\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.212362 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-config-data\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.212409 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-fernet-keys\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.314525 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-config-data\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.314583 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-fernet-keys\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.314724 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-combined-ca-bundle\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.314774 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw8wn\" (UniqueName: \"kubernetes.io/projected/46fd0927-55d9-4806-b7ad-2d3633aad62a-kube-api-access-xw8wn\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.320729 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-config-data\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.321125 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-fernet-keys\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.335688 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-combined-ca-bundle\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.336746 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw8wn\" (UniqueName: \"kubernetes.io/projected/46fd0927-55d9-4806-b7ad-2d3633aad62a-kube-api-access-xw8wn\") pod \"keystone-cron-29400301-dxpzw\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:00 crc kubenswrapper[5035]: I1124 21:01:00.478959 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:07 crc kubenswrapper[5035]: E1124 21:01:07.708123 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tobiko:current-podified" Nov 24 21:01:07 crc kubenswrapper[5035]: E1124 21:01:07.708961 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tobiko-tests-tobiko,Image:quay.io/podified-antelope-centos9/openstack-tobiko:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TOBIKO_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:TOBIKO_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:TOBIKO_LOGS_DIR_NAME,Value:tobiko-tests-tobiko-s00-podified-functional,ValueFrom:nil,},EnvVar{Name:TOBIKO_PYTEST_ADDOPTS,Value:,ValueFrom:nil,},EnvVar{Name:TOBIKO_TESTENV,Value:functional -- tobiko/tests/functional/podified/test_topology.py,ValueFrom:nil,},EnvVar{Name:TOBIKO_VERSION,Value:master,ValueFrom:nil,},EnvVar{Name:TOX_NUM_PROCESSES,Value:2,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{8 0} {} 8 DecimalSI},memory: {{8589934592 0} {} BinarySI},},Requests:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tobiko,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tobiko/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/tobiko/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-config,ReadOnly:false,MountPath:/etc/tobiko/tobiko.conf,SubPath:tobiko.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-private-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa,SubPath:id_ecdsa,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-public-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa.pub,SubPath:id_ecdsa.pub,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kubeconfig,ReadOnly:true,MountPath:/var/lib/tobiko/.kube/config,SubPath:config,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-smkwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42495,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42495,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tobiko-tests-tobiko-s00-podified-functional_openstack(3b0c51fa-ab78-4018-a361-f02d53b92345): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 21:01:07 crc kubenswrapper[5035]: E1124 21:01:07.710138 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="3b0c51fa-ab78-4018-a361-f02d53b92345" Nov 24 21:01:08 crc kubenswrapper[5035]: I1124 21:01:08.149231 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400301-dxpzw"] Nov 24 21:01:08 crc kubenswrapper[5035]: I1124 21:01:08.548987 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400301-dxpzw" event={"ID":"46fd0927-55d9-4806-b7ad-2d3633aad62a","Type":"ContainerStarted","Data":"f13fd6177dfcade10d10b9d03f5db4a1dd494e66723e8f82442f4572589c0193"} Nov 24 21:01:08 crc kubenswrapper[5035]: I1124 21:01:08.549347 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400301-dxpzw" event={"ID":"46fd0927-55d9-4806-b7ad-2d3633aad62a","Type":"ContainerStarted","Data":"6ebb0bbe12fb522774eaf46992d2bd8c7eca851deeaa5ce0b4e2bc536091c225"} Nov 24 21:01:08 crc kubenswrapper[5035]: E1124 21:01:08.552007 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tobiko:current-podified\\\"\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="3b0c51fa-ab78-4018-a361-f02d53b92345" Nov 24 21:01:08 crc kubenswrapper[5035]: I1124 21:01:08.574649 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29400301-dxpzw" podStartSLOduration=8.57462106 podStartE2EDuration="8.57462106s" podCreationTimestamp="2025-11-24 21:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 21:01:08.570981283 +0000 UTC m=+6407.093487570" watchObservedRunningTime="2025-11-24 21:01:08.57462106 +0000 UTC m=+6407.097127327" Nov 24 21:01:11 crc kubenswrapper[5035]: I1124 21:01:11.582732 5035 generic.go:334] "Generic (PLEG): container finished" podID="46fd0927-55d9-4806-b7ad-2d3633aad62a" containerID="f13fd6177dfcade10d10b9d03f5db4a1dd494e66723e8f82442f4572589c0193" exitCode=0 Nov 24 21:01:11 crc kubenswrapper[5035]: I1124 21:01:11.582891 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400301-dxpzw" event={"ID":"46fd0927-55d9-4806-b7ad-2d3633aad62a","Type":"ContainerDied","Data":"f13fd6177dfcade10d10b9d03f5db4a1dd494e66723e8f82442f4572589c0193"} Nov 24 21:01:12 crc kubenswrapper[5035]: I1124 21:01:12.981739 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.044579 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw8wn\" (UniqueName: \"kubernetes.io/projected/46fd0927-55d9-4806-b7ad-2d3633aad62a-kube-api-access-xw8wn\") pod \"46fd0927-55d9-4806-b7ad-2d3633aad62a\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.044768 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-combined-ca-bundle\") pod \"46fd0927-55d9-4806-b7ad-2d3633aad62a\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.045030 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-config-data\") pod \"46fd0927-55d9-4806-b7ad-2d3633aad62a\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.045097 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-fernet-keys\") pod \"46fd0927-55d9-4806-b7ad-2d3633aad62a\" (UID: \"46fd0927-55d9-4806-b7ad-2d3633aad62a\") " Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.051882 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "46fd0927-55d9-4806-b7ad-2d3633aad62a" (UID: "46fd0927-55d9-4806-b7ad-2d3633aad62a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.053627 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46fd0927-55d9-4806-b7ad-2d3633aad62a-kube-api-access-xw8wn" (OuterVolumeSpecName: "kube-api-access-xw8wn") pod "46fd0927-55d9-4806-b7ad-2d3633aad62a" (UID: "46fd0927-55d9-4806-b7ad-2d3633aad62a"). InnerVolumeSpecName "kube-api-access-xw8wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.072807 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46fd0927-55d9-4806-b7ad-2d3633aad62a" (UID: "46fd0927-55d9-4806-b7ad-2d3633aad62a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.114611 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-config-data" (OuterVolumeSpecName: "config-data") pod "46fd0927-55d9-4806-b7ad-2d3633aad62a" (UID: "46fd0927-55d9-4806-b7ad-2d3633aad62a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.148152 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw8wn\" (UniqueName: \"kubernetes.io/projected/46fd0927-55d9-4806-b7ad-2d3633aad62a-kube-api-access-xw8wn\") on node \"crc\" DevicePath \"\"" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.148190 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.148203 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.148219 5035 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/46fd0927-55d9-4806-b7ad-2d3633aad62a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.608170 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400301-dxpzw" event={"ID":"46fd0927-55d9-4806-b7ad-2d3633aad62a","Type":"ContainerDied","Data":"6ebb0bbe12fb522774eaf46992d2bd8c7eca851deeaa5ce0b4e2bc536091c225"} Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.608243 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ebb0bbe12fb522774eaf46992d2bd8c7eca851deeaa5ce0b4e2bc536091c225" Nov 24 21:01:13 crc kubenswrapper[5035]: I1124 21:01:13.608434 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400301-dxpzw" Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.234606 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.235017 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.235104 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.236453 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.236612 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" gracePeriod=600 Nov 24 21:01:15 crc kubenswrapper[5035]: E1124 21:01:15.371417 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.637247 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" exitCode=0 Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.637398 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3"} Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.637677 5035 scope.go:117] "RemoveContainer" containerID="ec02eac9e2bb0061a2c002a2e8f9ba07191c078cfe0a44c8ad6165a85a41d8a2" Nov 24 21:01:15 crc kubenswrapper[5035]: I1124 21:01:15.638680 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:01:15 crc kubenswrapper[5035]: E1124 21:01:15.639199 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:01:24 crc kubenswrapper[5035]: I1124 21:01:24.756433 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"3b0c51fa-ab78-4018-a361-f02d53b92345","Type":"ContainerStarted","Data":"bb6de1f856369182dcd72d0e3ddc5ca139abba1cf606f50f04ce0f6c46a63870"} Nov 24 21:01:24 crc kubenswrapper[5035]: I1124 21:01:24.784890 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podStartSLOduration=3.418326214 podStartE2EDuration="38.784871124s" podCreationTimestamp="2025-11-24 21:00:46 +0000 UTC" firstStartedPulling="2025-11-24 21:00:48.335046657 +0000 UTC m=+6386.857552914" lastFinishedPulling="2025-11-24 21:01:23.701591527 +0000 UTC m=+6422.224097824" observedRunningTime="2025-11-24 21:01:24.777327602 +0000 UTC m=+6423.299833869" watchObservedRunningTime="2025-11-24 21:01:24.784871124 +0000 UTC m=+6423.307377381" Nov 24 21:01:28 crc kubenswrapper[5035]: I1124 21:01:28.200569 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:01:28 crc kubenswrapper[5035]: E1124 21:01:28.201849 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:01:43 crc kubenswrapper[5035]: I1124 21:01:43.201000 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:01:43 crc kubenswrapper[5035]: E1124 21:01:43.201740 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:01:54 crc kubenswrapper[5035]: E1124 21:01:54.782539 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Nov 24 21:01:58 crc kubenswrapper[5035]: I1124 21:01:58.200735 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:01:58 crc kubenswrapper[5035]: E1124 21:01:58.201515 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:02:09 crc kubenswrapper[5035]: I1124 21:02:09.200862 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:02:09 crc kubenswrapper[5035]: E1124 21:02:09.202070 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:02:21 crc kubenswrapper[5035]: I1124 21:02:21.200707 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:02:21 crc kubenswrapper[5035]: E1124 21:02:21.202261 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:02:30 crc kubenswrapper[5035]: I1124 21:02:30.642416 5035 generic.go:334] "Generic (PLEG): container finished" podID="3b0c51fa-ab78-4018-a361-f02d53b92345" containerID="bb6de1f856369182dcd72d0e3ddc5ca139abba1cf606f50f04ce0f6c46a63870" exitCode=0 Nov 24 21:02:30 crc kubenswrapper[5035]: I1124 21:02:30.642535 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"3b0c51fa-ab78-4018-a361-f02d53b92345","Type":"ContainerDied","Data":"bb6de1f856369182dcd72d0e3ddc5ca139abba1cf606f50f04ce0f6c46a63870"} Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.158988 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.251511 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:02:32 crc kubenswrapper[5035]: E1124 21:02:32.251764 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.257176 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Nov 24 21:02:32 crc kubenswrapper[5035]: E1124 21:02:32.257697 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0c51fa-ab78-4018-a361-f02d53b92345" containerName="tobiko-tests-tobiko" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.257723 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0c51fa-ab78-4018-a361-f02d53b92345" containerName="tobiko-tests-tobiko" Nov 24 21:02:32 crc kubenswrapper[5035]: E1124 21:02:32.257760 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46fd0927-55d9-4806-b7ad-2d3633aad62a" containerName="keystone-cron" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.257770 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="46fd0927-55d9-4806-b7ad-2d3633aad62a" containerName="keystone-cron" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.258042 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="46fd0927-55d9-4806-b7ad-2d3633aad62a" containerName="keystone-cron" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.258068 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0c51fa-ab78-4018-a361-f02d53b92345" containerName="tobiko-tests-tobiko" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.258902 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.266849 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.268797 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-kubeconfig\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.269232 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-temporary\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.269473 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-clouds-config\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.269611 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-private-key\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.269714 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-openstack-config-secret\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.269838 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smkwb\" (UniqueName: \"kubernetes.io/projected/3b0c51fa-ab78-4018-a361-f02d53b92345-kube-api-access-smkwb\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.269926 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ceph\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.270038 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ca-certs\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.270181 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-workdir\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.270312 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-public-key\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.270461 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.270703 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-config\") pod \"3b0c51fa-ab78-4018-a361-f02d53b92345\" (UID: \"3b0c51fa-ab78-4018-a361-f02d53b92345\") " Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.270907 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.271474 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.286078 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ceph" (OuterVolumeSpecName: "ceph") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.293872 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.293919 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0c51fa-ab78-4018-a361-f02d53b92345-kube-api-access-smkwb" (OuterVolumeSpecName: "kube-api-access-smkwb") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "kube-api-access-smkwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.299220 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.300976 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.306444 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.321331 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.322586 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.327235 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.343906 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373668 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373737 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373780 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373824 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373863 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373878 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373915 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373934 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxprm\" (UniqueName: \"kubernetes.io/projected/7af30223-20cb-4470-b630-170318d5de9e-kube-api-access-xxprm\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.373983 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374043 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374079 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374106 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374201 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374218 5035 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374231 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374245 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smkwb\" (UniqueName: \"kubernetes.io/projected/3b0c51fa-ab78-4018-a361-f02d53b92345-kube-api-access-smkwb\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374258 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374271 5035 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374282 5035 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374315 5035 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3b0c51fa-ab78-4018-a361-f02d53b92345-tobiko-config\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.374326 5035 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3b0c51fa-ab78-4018-a361-f02d53b92345-kubeconfig\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.398584 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475671 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475742 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475794 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475817 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475856 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxprm\" (UniqueName: \"kubernetes.io/projected/7af30223-20cb-4470-b630-170318d5de9e-kube-api-access-xxprm\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475898 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475932 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475957 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.475984 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.476047 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.476087 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.476557 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.476974 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.478242 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.478264 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.479647 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.479685 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.480212 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.481964 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.482063 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.484573 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.492218 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxprm\" (UniqueName: \"kubernetes.io/projected/7af30223-20cb-4470-b630-170318d5de9e-kube-api-access-xxprm\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.588798 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.667389 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"3b0c51fa-ab78-4018-a361-f02d53b92345","Type":"ContainerDied","Data":"648f05ec6b3c8cfc57b50c1454d8e6fb55927443fa2032d9c96d710a5e489ed2"} Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.667427 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="648f05ec6b3c8cfc57b50c1454d8e6fb55927443fa2032d9c96d710a5e489ed2" Nov 24 21:02:32 crc kubenswrapper[5035]: I1124 21:02:32.667675 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Nov 24 21:02:33 crc kubenswrapper[5035]: I1124 21:02:33.229165 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Nov 24 21:02:33 crc kubenswrapper[5035]: I1124 21:02:33.685872 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"7af30223-20cb-4470-b630-170318d5de9e","Type":"ContainerStarted","Data":"8ebbe115cb8b60416502ccdc02a01c49fdc11b7e69a9089d77b27546fe8e9d6c"} Nov 24 21:02:33 crc kubenswrapper[5035]: I1124 21:02:33.761390 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3b0c51fa-ab78-4018-a361-f02d53b92345" (UID: "3b0c51fa-ab78-4018-a361-f02d53b92345"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:02:33 crc kubenswrapper[5035]: I1124 21:02:33.806310 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b0c51fa-ab78-4018-a361-f02d53b92345-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:34 crc kubenswrapper[5035]: I1124 21:02:34.703732 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"7af30223-20cb-4470-b630-170318d5de9e","Type":"ContainerStarted","Data":"1d61a498c3f890146ab9b4327269b4408b8c686db8561eceac4eb014353b691a"} Nov 24 21:02:34 crc kubenswrapper[5035]: I1124 21:02:34.724348 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s01-sanity" podStartSLOduration=2.7242722390000003 podStartE2EDuration="2.724272239s" podCreationTimestamp="2025-11-24 21:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 21:02:34.72165983 +0000 UTC m=+6493.244166137" watchObservedRunningTime="2025-11-24 21:02:34.724272239 +0000 UTC m=+6493.246778546" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.577154 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q2k7j"] Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.581208 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.606039 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2k7j"] Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.754882 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-utilities\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.755283 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6ng9\" (UniqueName: \"kubernetes.io/projected/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-kube-api-access-c6ng9\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.755319 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-catalog-content\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.857362 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-utilities\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.857516 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6ng9\" (UniqueName: \"kubernetes.io/projected/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-kube-api-access-c6ng9\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.857544 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-catalog-content\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.858226 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-utilities\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.858260 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-catalog-content\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.879138 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6ng9\" (UniqueName: \"kubernetes.io/projected/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-kube-api-access-c6ng9\") pod \"certified-operators-q2k7j\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:43 crc kubenswrapper[5035]: I1124 21:02:43.903032 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:44 crc kubenswrapper[5035]: I1124 21:02:44.449544 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2k7j"] Nov 24 21:02:44 crc kubenswrapper[5035]: W1124 21:02:44.462731 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cf043a2_ae2f_4332_a1d0_e034c2f6b9b7.slice/crio-71acd27a75730e30533d3fb20551a999d25096b80e6484f0e71e37d4ac952de6 WatchSource:0}: Error finding container 71acd27a75730e30533d3fb20551a999d25096b80e6484f0e71e37d4ac952de6: Status 404 returned error can't find the container with id 71acd27a75730e30533d3fb20551a999d25096b80e6484f0e71e37d4ac952de6 Nov 24 21:02:44 crc kubenswrapper[5035]: I1124 21:02:44.823644 5035 generic.go:334] "Generic (PLEG): container finished" podID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerID="da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb" exitCode=0 Nov 24 21:02:44 crc kubenswrapper[5035]: I1124 21:02:44.823952 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2k7j" event={"ID":"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7","Type":"ContainerDied","Data":"da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb"} Nov 24 21:02:44 crc kubenswrapper[5035]: I1124 21:02:44.824013 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2k7j" event={"ID":"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7","Type":"ContainerStarted","Data":"71acd27a75730e30533d3fb20551a999d25096b80e6484f0e71e37d4ac952de6"} Nov 24 21:02:45 crc kubenswrapper[5035]: I1124 21:02:45.200998 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:02:45 crc kubenswrapper[5035]: E1124 21:02:45.201450 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:02:46 crc kubenswrapper[5035]: I1124 21:02:46.852889 5035 generic.go:334] "Generic (PLEG): container finished" podID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerID="e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae" exitCode=0 Nov 24 21:02:46 crc kubenswrapper[5035]: I1124 21:02:46.852982 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2k7j" event={"ID":"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7","Type":"ContainerDied","Data":"e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae"} Nov 24 21:02:47 crc kubenswrapper[5035]: I1124 21:02:47.865818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2k7j" event={"ID":"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7","Type":"ContainerStarted","Data":"70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7"} Nov 24 21:02:47 crc kubenswrapper[5035]: I1124 21:02:47.895936 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q2k7j" podStartSLOduration=2.439593649 podStartE2EDuration="4.895917457s" podCreationTimestamp="2025-11-24 21:02:43 +0000 UTC" firstStartedPulling="2025-11-24 21:02:44.825506359 +0000 UTC m=+6503.348012616" lastFinishedPulling="2025-11-24 21:02:47.281830157 +0000 UTC m=+6505.804336424" observedRunningTime="2025-11-24 21:02:47.887619475 +0000 UTC m=+6506.410125752" watchObservedRunningTime="2025-11-24 21:02:47.895917457 +0000 UTC m=+6506.418423714" Nov 24 21:02:53 crc kubenswrapper[5035]: I1124 21:02:53.903274 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:53 crc kubenswrapper[5035]: I1124 21:02:53.903908 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:54 crc kubenswrapper[5035]: I1124 21:02:54.002069 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:54 crc kubenswrapper[5035]: I1124 21:02:54.081408 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:54 crc kubenswrapper[5035]: I1124 21:02:54.247891 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2k7j"] Nov 24 21:02:55 crc kubenswrapper[5035]: I1124 21:02:55.963211 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q2k7j" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="registry-server" containerID="cri-o://70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7" gracePeriod=2 Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.498023 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.552884 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-catalog-content\") pod \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.654828 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6ng9\" (UniqueName: \"kubernetes.io/projected/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-kube-api-access-c6ng9\") pod \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.655201 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-utilities\") pod \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\" (UID: \"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7\") " Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.656244 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-utilities" (OuterVolumeSpecName: "utilities") pod "1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" (UID: "1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.675277 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-kube-api-access-c6ng9" (OuterVolumeSpecName: "kube-api-access-c6ng9") pod "1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" (UID: "1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7"). InnerVolumeSpecName "kube-api-access-c6ng9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.758071 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.758142 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6ng9\" (UniqueName: \"kubernetes.io/projected/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-kube-api-access-c6ng9\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.765566 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" (UID: "1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.862388 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.980490 5035 generic.go:334] "Generic (PLEG): container finished" podID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerID="70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7" exitCode=0 Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.980557 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2k7j" event={"ID":"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7","Type":"ContainerDied","Data":"70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7"} Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.980598 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2k7j" event={"ID":"1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7","Type":"ContainerDied","Data":"71acd27a75730e30533d3fb20551a999d25096b80e6484f0e71e37d4ac952de6"} Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.980631 5035 scope.go:117] "RemoveContainer" containerID="70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7" Nov 24 21:02:56 crc kubenswrapper[5035]: I1124 21:02:56.980618 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2k7j" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.014802 5035 scope.go:117] "RemoveContainer" containerID="e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.042841 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q2k7j"] Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.054677 5035 scope.go:117] "RemoveContainer" containerID="da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.057189 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q2k7j"] Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.114730 5035 scope.go:117] "RemoveContainer" containerID="70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7" Nov 24 21:02:57 crc kubenswrapper[5035]: E1124 21:02:57.115144 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7\": container with ID starting with 70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7 not found: ID does not exist" containerID="70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.115183 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7"} err="failed to get container status \"70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7\": rpc error: code = NotFound desc = could not find container \"70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7\": container with ID starting with 70061c9614e6a82436993594a5083131b9bab57a88e8e095e611098d4c4973e7 not found: ID does not exist" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.115211 5035 scope.go:117] "RemoveContainer" containerID="e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae" Nov 24 21:02:57 crc kubenswrapper[5035]: E1124 21:02:57.115482 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae\": container with ID starting with e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae not found: ID does not exist" containerID="e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.115518 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae"} err="failed to get container status \"e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae\": rpc error: code = NotFound desc = could not find container \"e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae\": container with ID starting with e30498b207ef38252dbc2b472eec6773135ec962aad3017d8ab6474a12a64aae not found: ID does not exist" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.115538 5035 scope.go:117] "RemoveContainer" containerID="da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb" Nov 24 21:02:57 crc kubenswrapper[5035]: E1124 21:02:57.115761 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb\": container with ID starting with da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb not found: ID does not exist" containerID="da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.115791 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb"} err="failed to get container status \"da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb\": rpc error: code = NotFound desc = could not find container \"da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb\": container with ID starting with da5b499248d5e227fc2b3482b053b1f95ba7143ad26c3a0e69bbe7ef1ca8c4fb not found: ID does not exist" Nov 24 21:02:57 crc kubenswrapper[5035]: I1124 21:02:57.200906 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:02:57 crc kubenswrapper[5035]: E1124 21:02:57.201325 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:02:58 crc kubenswrapper[5035]: I1124 21:02:58.218399 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" path="/var/lib/kubelet/pods/1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7/volumes" Nov 24 21:03:11 crc kubenswrapper[5035]: I1124 21:03:11.200037 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:03:11 crc kubenswrapper[5035]: E1124 21:03:11.201158 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:03:25 crc kubenswrapper[5035]: I1124 21:03:25.199959 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:03:25 crc kubenswrapper[5035]: E1124 21:03:25.201118 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:03:39 crc kubenswrapper[5035]: I1124 21:03:39.199947 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:03:39 crc kubenswrapper[5035]: E1124 21:03:39.200840 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:03:53 crc kubenswrapper[5035]: I1124 21:03:53.201038 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:03:53 crc kubenswrapper[5035]: E1124 21:03:53.202378 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.778370 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mtkj6"] Nov 24 21:03:54 crc kubenswrapper[5035]: E1124 21:03:54.779118 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="registry-server" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.779134 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="registry-server" Nov 24 21:03:54 crc kubenswrapper[5035]: E1124 21:03:54.779163 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="extract-utilities" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.779173 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="extract-utilities" Nov 24 21:03:54 crc kubenswrapper[5035]: E1124 21:03:54.779182 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="extract-content" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.779190 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="extract-content" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.779440 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cf043a2-ae2f-4332-a1d0-e034c2f6b9b7" containerName="registry-server" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.813550 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtkj6"] Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.813700 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.868830 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk2cr\" (UniqueName: \"kubernetes.io/projected/275427ec-b4f3-4e6f-9a16-e77068388dd5-kube-api-access-qk2cr\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.868911 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-utilities\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.869035 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-catalog-content\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.970884 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk2cr\" (UniqueName: \"kubernetes.io/projected/275427ec-b4f3-4e6f-9a16-e77068388dd5-kube-api-access-qk2cr\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.970946 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-utilities\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.971034 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-catalog-content\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.971504 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-catalog-content\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.971939 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-utilities\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:54 crc kubenswrapper[5035]: I1124 21:03:54.998059 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk2cr\" (UniqueName: \"kubernetes.io/projected/275427ec-b4f3-4e6f-9a16-e77068388dd5-kube-api-access-qk2cr\") pod \"redhat-marketplace-mtkj6\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:55 crc kubenswrapper[5035]: I1124 21:03:55.144010 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:03:55 crc kubenswrapper[5035]: I1124 21:03:55.667489 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtkj6"] Nov 24 21:03:55 crc kubenswrapper[5035]: W1124 21:03:55.677565 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod275427ec_b4f3_4e6f_9a16_e77068388dd5.slice/crio-643f8a85a5fa231c76edf2ec740acdc19f1d69e3c1d1a5ec792c40d60ec30206 WatchSource:0}: Error finding container 643f8a85a5fa231c76edf2ec740acdc19f1d69e3c1d1a5ec792c40d60ec30206: Status 404 returned error can't find the container with id 643f8a85a5fa231c76edf2ec740acdc19f1d69e3c1d1a5ec792c40d60ec30206 Nov 24 21:03:56 crc kubenswrapper[5035]: I1124 21:03:56.594246 5035 generic.go:334] "Generic (PLEG): container finished" podID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerID="31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6" exitCode=0 Nov 24 21:03:56 crc kubenswrapper[5035]: I1124 21:03:56.594366 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtkj6" event={"ID":"275427ec-b4f3-4e6f-9a16-e77068388dd5","Type":"ContainerDied","Data":"31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6"} Nov 24 21:03:56 crc kubenswrapper[5035]: I1124 21:03:56.594751 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtkj6" event={"ID":"275427ec-b4f3-4e6f-9a16-e77068388dd5","Type":"ContainerStarted","Data":"643f8a85a5fa231c76edf2ec740acdc19f1d69e3c1d1a5ec792c40d60ec30206"} Nov 24 21:03:56 crc kubenswrapper[5035]: I1124 21:03:56.599577 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 21:03:57 crc kubenswrapper[5035]: I1124 21:03:57.606523 5035 generic.go:334] "Generic (PLEG): container finished" podID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerID="859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9" exitCode=0 Nov 24 21:03:57 crc kubenswrapper[5035]: I1124 21:03:57.606795 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtkj6" event={"ID":"275427ec-b4f3-4e6f-9a16-e77068388dd5","Type":"ContainerDied","Data":"859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9"} Nov 24 21:03:58 crc kubenswrapper[5035]: I1124 21:03:58.620135 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtkj6" event={"ID":"275427ec-b4f3-4e6f-9a16-e77068388dd5","Type":"ContainerStarted","Data":"64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9"} Nov 24 21:03:58 crc kubenswrapper[5035]: I1124 21:03:58.650772 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mtkj6" podStartSLOduration=3.176979799 podStartE2EDuration="4.65073142s" podCreationTimestamp="2025-11-24 21:03:54 +0000 UTC" firstStartedPulling="2025-11-24 21:03:56.599325625 +0000 UTC m=+6575.121831882" lastFinishedPulling="2025-11-24 21:03:58.073077206 +0000 UTC m=+6576.595583503" observedRunningTime="2025-11-24 21:03:58.644722618 +0000 UTC m=+6577.167228915" watchObservedRunningTime="2025-11-24 21:03:58.65073142 +0000 UTC m=+6577.173237687" Nov 24 21:04:04 crc kubenswrapper[5035]: I1124 21:04:04.200826 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:04:04 crc kubenswrapper[5035]: E1124 21:04:04.202615 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:04:05 crc kubenswrapper[5035]: I1124 21:04:05.144849 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:04:05 crc kubenswrapper[5035]: I1124 21:04:05.145966 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:04:05 crc kubenswrapper[5035]: I1124 21:04:05.196009 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:04:05 crc kubenswrapper[5035]: I1124 21:04:05.770076 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:04:05 crc kubenswrapper[5035]: I1124 21:04:05.841365 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtkj6"] Nov 24 21:04:07 crc kubenswrapper[5035]: I1124 21:04:07.718089 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mtkj6" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="registry-server" containerID="cri-o://64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9" gracePeriod=2 Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.248658 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.268973 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-catalog-content\") pod \"275427ec-b4f3-4e6f-9a16-e77068388dd5\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.269048 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk2cr\" (UniqueName: \"kubernetes.io/projected/275427ec-b4f3-4e6f-9a16-e77068388dd5-kube-api-access-qk2cr\") pod \"275427ec-b4f3-4e6f-9a16-e77068388dd5\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.269202 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-utilities\") pod \"275427ec-b4f3-4e6f-9a16-e77068388dd5\" (UID: \"275427ec-b4f3-4e6f-9a16-e77068388dd5\") " Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.273389 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-utilities" (OuterVolumeSpecName: "utilities") pod "275427ec-b4f3-4e6f-9a16-e77068388dd5" (UID: "275427ec-b4f3-4e6f-9a16-e77068388dd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.291449 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/275427ec-b4f3-4e6f-9a16-e77068388dd5-kube-api-access-qk2cr" (OuterVolumeSpecName: "kube-api-access-qk2cr") pod "275427ec-b4f3-4e6f-9a16-e77068388dd5" (UID: "275427ec-b4f3-4e6f-9a16-e77068388dd5"). InnerVolumeSpecName "kube-api-access-qk2cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.305610 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "275427ec-b4f3-4e6f-9a16-e77068388dd5" (UID: "275427ec-b4f3-4e6f-9a16-e77068388dd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.371398 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.371430 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk2cr\" (UniqueName: \"kubernetes.io/projected/275427ec-b4f3-4e6f-9a16-e77068388dd5-kube-api-access-qk2cr\") on node \"crc\" DevicePath \"\"" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.371439 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275427ec-b4f3-4e6f-9a16-e77068388dd5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.737273 5035 generic.go:334] "Generic (PLEG): container finished" podID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerID="64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9" exitCode=0 Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.737566 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtkj6" event={"ID":"275427ec-b4f3-4e6f-9a16-e77068388dd5","Type":"ContainerDied","Data":"64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9"} Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.737758 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtkj6" event={"ID":"275427ec-b4f3-4e6f-9a16-e77068388dd5","Type":"ContainerDied","Data":"643f8a85a5fa231c76edf2ec740acdc19f1d69e3c1d1a5ec792c40d60ec30206"} Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.737674 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtkj6" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.737810 5035 scope.go:117] "RemoveContainer" containerID="64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.771686 5035 scope.go:117] "RemoveContainer" containerID="859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.795772 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtkj6"] Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.809532 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtkj6"] Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.813222 5035 scope.go:117] "RemoveContainer" containerID="31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.868556 5035 scope.go:117] "RemoveContainer" containerID="64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9" Nov 24 21:04:08 crc kubenswrapper[5035]: E1124 21:04:08.869023 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9\": container with ID starting with 64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9 not found: ID does not exist" containerID="64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.869082 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9"} err="failed to get container status \"64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9\": rpc error: code = NotFound desc = could not find container \"64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9\": container with ID starting with 64875cf5ba026a5e19d949620602287603c6ffee31573a9673eb5e69766899e9 not found: ID does not exist" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.869125 5035 scope.go:117] "RemoveContainer" containerID="859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9" Nov 24 21:04:08 crc kubenswrapper[5035]: E1124 21:04:08.869645 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9\": container with ID starting with 859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9 not found: ID does not exist" containerID="859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.869686 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9"} err="failed to get container status \"859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9\": rpc error: code = NotFound desc = could not find container \"859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9\": container with ID starting with 859de71210bf591d66e7a38fb7c79aaf47112c44fde034cce2ea0d349f9994a9 not found: ID does not exist" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.869712 5035 scope.go:117] "RemoveContainer" containerID="31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6" Nov 24 21:04:08 crc kubenswrapper[5035]: E1124 21:04:08.870008 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6\": container with ID starting with 31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6 not found: ID does not exist" containerID="31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6" Nov 24 21:04:08 crc kubenswrapper[5035]: I1124 21:04:08.870054 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6"} err="failed to get container status \"31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6\": rpc error: code = NotFound desc = could not find container \"31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6\": container with ID starting with 31e40f446db5a7d6217f1cb828d29f5349a86553c0b22260b1ccf89e346644d6 not found: ID does not exist" Nov 24 21:04:10 crc kubenswrapper[5035]: I1124 21:04:10.219022 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" path="/var/lib/kubelet/pods/275427ec-b4f3-4e6f-9a16-e77068388dd5/volumes" Nov 24 21:04:19 crc kubenswrapper[5035]: I1124 21:04:19.201411 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:04:19 crc kubenswrapper[5035]: E1124 21:04:19.204268 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:04:22 crc kubenswrapper[5035]: I1124 21:04:22.830621 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4hssh"] Nov 24 21:04:22 crc kubenswrapper[5035]: E1124 21:04:22.831758 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="extract-content" Nov 24 21:04:22 crc kubenswrapper[5035]: I1124 21:04:22.831782 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="extract-content" Nov 24 21:04:22 crc kubenswrapper[5035]: E1124 21:04:22.831821 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="extract-utilities" Nov 24 21:04:22 crc kubenswrapper[5035]: I1124 21:04:22.831832 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="extract-utilities" Nov 24 21:04:22 crc kubenswrapper[5035]: E1124 21:04:22.831856 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="registry-server" Nov 24 21:04:22 crc kubenswrapper[5035]: I1124 21:04:22.831867 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="registry-server" Nov 24 21:04:22 crc kubenswrapper[5035]: I1124 21:04:22.832189 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="275427ec-b4f3-4e6f-9a16-e77068388dd5" containerName="registry-server" Nov 24 21:04:22 crc kubenswrapper[5035]: I1124 21:04:22.842489 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:22 crc kubenswrapper[5035]: I1124 21:04:22.857893 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4hssh"] Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.004008 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2v7m\" (UniqueName: \"kubernetes.io/projected/bcc5984b-4d6f-4927-86c9-4f3711183e1c-kube-api-access-c2v7m\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.004407 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-catalog-content\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.004432 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-utilities\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.106438 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2v7m\" (UniqueName: \"kubernetes.io/projected/bcc5984b-4d6f-4927-86c9-4f3711183e1c-kube-api-access-c2v7m\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.106515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-catalog-content\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.106544 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-utilities\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.107204 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-utilities\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.107432 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-catalog-content\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.134662 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2v7m\" (UniqueName: \"kubernetes.io/projected/bcc5984b-4d6f-4927-86c9-4f3711183e1c-kube-api-access-c2v7m\") pod \"redhat-operators-4hssh\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.186880 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.705708 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4hssh"] Nov 24 21:04:23 crc kubenswrapper[5035]: I1124 21:04:23.921515 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hssh" event={"ID":"bcc5984b-4d6f-4927-86c9-4f3711183e1c","Type":"ContainerStarted","Data":"0f86620bae60c084f5601fc7ce709157bd1a5c2031d101c1c30d407afff1a44a"} Nov 24 21:04:24 crc kubenswrapper[5035]: I1124 21:04:24.936822 5035 generic.go:334] "Generic (PLEG): container finished" podID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerID="81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724" exitCode=0 Nov 24 21:04:24 crc kubenswrapper[5035]: I1124 21:04:24.936877 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hssh" event={"ID":"bcc5984b-4d6f-4927-86c9-4f3711183e1c","Type":"ContainerDied","Data":"81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724"} Nov 24 21:04:25 crc kubenswrapper[5035]: I1124 21:04:25.950489 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hssh" event={"ID":"bcc5984b-4d6f-4927-86c9-4f3711183e1c","Type":"ContainerStarted","Data":"6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716"} Nov 24 21:04:26 crc kubenswrapper[5035]: I1124 21:04:26.961661 5035 generic.go:334] "Generic (PLEG): container finished" podID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerID="6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716" exitCode=0 Nov 24 21:04:26 crc kubenswrapper[5035]: I1124 21:04:26.961708 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hssh" event={"ID":"bcc5984b-4d6f-4927-86c9-4f3711183e1c","Type":"ContainerDied","Data":"6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716"} Nov 24 21:04:27 crc kubenswrapper[5035]: I1124 21:04:27.978069 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hssh" event={"ID":"bcc5984b-4d6f-4927-86c9-4f3711183e1c","Type":"ContainerStarted","Data":"03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381"} Nov 24 21:04:28 crc kubenswrapper[5035]: I1124 21:04:28.015377 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4hssh" podStartSLOduration=3.353668873 podStartE2EDuration="6.015351805s" podCreationTimestamp="2025-11-24 21:04:22 +0000 UTC" firstStartedPulling="2025-11-24 21:04:24.939662957 +0000 UTC m=+6603.462169254" lastFinishedPulling="2025-11-24 21:04:27.601345919 +0000 UTC m=+6606.123852186" observedRunningTime="2025-11-24 21:04:28.000892728 +0000 UTC m=+6606.523399065" watchObservedRunningTime="2025-11-24 21:04:28.015351805 +0000 UTC m=+6606.537858102" Nov 24 21:04:32 crc kubenswrapper[5035]: I1124 21:04:32.206061 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:04:32 crc kubenswrapper[5035]: E1124 21:04:32.206790 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:04:33 crc kubenswrapper[5035]: I1124 21:04:33.191575 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:33 crc kubenswrapper[5035]: I1124 21:04:33.192278 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:34 crc kubenswrapper[5035]: I1124 21:04:34.260187 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4hssh" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="registry-server" probeResult="failure" output=< Nov 24 21:04:34 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 21:04:34 crc kubenswrapper[5035]: > Nov 24 21:04:43 crc kubenswrapper[5035]: I1124 21:04:43.246689 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:43 crc kubenswrapper[5035]: I1124 21:04:43.299454 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:43 crc kubenswrapper[5035]: I1124 21:04:43.486640 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4hssh"] Nov 24 21:04:44 crc kubenswrapper[5035]: I1124 21:04:44.201639 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:04:44 crc kubenswrapper[5035]: E1124 21:04:44.201868 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.157595 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4hssh" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="registry-server" containerID="cri-o://03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381" gracePeriod=2 Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.658544 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.816660 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-catalog-content\") pod \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.816793 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2v7m\" (UniqueName: \"kubernetes.io/projected/bcc5984b-4d6f-4927-86c9-4f3711183e1c-kube-api-access-c2v7m\") pod \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.816903 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-utilities\") pod \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\" (UID: \"bcc5984b-4d6f-4927-86c9-4f3711183e1c\") " Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.819521 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-utilities" (OuterVolumeSpecName: "utilities") pod "bcc5984b-4d6f-4927-86c9-4f3711183e1c" (UID: "bcc5984b-4d6f-4927-86c9-4f3711183e1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.827670 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc5984b-4d6f-4927-86c9-4f3711183e1c-kube-api-access-c2v7m" (OuterVolumeSpecName: "kube-api-access-c2v7m") pod "bcc5984b-4d6f-4927-86c9-4f3711183e1c" (UID: "bcc5984b-4d6f-4927-86c9-4f3711183e1c"). InnerVolumeSpecName "kube-api-access-c2v7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.920047 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2v7m\" (UniqueName: \"kubernetes.io/projected/bcc5984b-4d6f-4927-86c9-4f3711183e1c-kube-api-access-c2v7m\") on node \"crc\" DevicePath \"\"" Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.920096 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:04:45 crc kubenswrapper[5035]: I1124 21:04:45.942059 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcc5984b-4d6f-4927-86c9-4f3711183e1c" (UID: "bcc5984b-4d6f-4927-86c9-4f3711183e1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.021965 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc5984b-4d6f-4927-86c9-4f3711183e1c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.168309 5035 generic.go:334] "Generic (PLEG): container finished" podID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerID="03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381" exitCode=0 Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.168367 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hssh" event={"ID":"bcc5984b-4d6f-4927-86c9-4f3711183e1c","Type":"ContainerDied","Data":"03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381"} Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.168411 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4hssh" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.168434 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4hssh" event={"ID":"bcc5984b-4d6f-4927-86c9-4f3711183e1c","Type":"ContainerDied","Data":"0f86620bae60c084f5601fc7ce709157bd1a5c2031d101c1c30d407afff1a44a"} Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.168465 5035 scope.go:117] "RemoveContainer" containerID="03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.204282 5035 scope.go:117] "RemoveContainer" containerID="6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.216839 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4hssh"] Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.225030 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4hssh"] Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.264144 5035 scope.go:117] "RemoveContainer" containerID="81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.300475 5035 scope.go:117] "RemoveContainer" containerID="03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381" Nov 24 21:04:46 crc kubenswrapper[5035]: E1124 21:04:46.301031 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381\": container with ID starting with 03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381 not found: ID does not exist" containerID="03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.301073 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381"} err="failed to get container status \"03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381\": rpc error: code = NotFound desc = could not find container \"03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381\": container with ID starting with 03e54971c3efa4dfa44e0caf71966c1de529a7b31e12966c90dd0716a092c381 not found: ID does not exist" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.301098 5035 scope.go:117] "RemoveContainer" containerID="6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716" Nov 24 21:04:46 crc kubenswrapper[5035]: E1124 21:04:46.301532 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716\": container with ID starting with 6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716 not found: ID does not exist" containerID="6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.301602 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716"} err="failed to get container status \"6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716\": rpc error: code = NotFound desc = could not find container \"6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716\": container with ID starting with 6162aceab7013feedb56964ec9935ab38d193cac5e1cfb4fe1b8a11f8530b716 not found: ID does not exist" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.301647 5035 scope.go:117] "RemoveContainer" containerID="81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724" Nov 24 21:04:46 crc kubenswrapper[5035]: E1124 21:04:46.302113 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724\": container with ID starting with 81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724 not found: ID does not exist" containerID="81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724" Nov 24 21:04:46 crc kubenswrapper[5035]: I1124 21:04:46.302171 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724"} err="failed to get container status \"81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724\": rpc error: code = NotFound desc = could not find container \"81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724\": container with ID starting with 81507092cc770ab11cca0151293ea4db95c857a9171640b736acb3e779bda724 not found: ID does not exist" Nov 24 21:04:48 crc kubenswrapper[5035]: I1124 21:04:48.210688 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" path="/var/lib/kubelet/pods/bcc5984b-4d6f-4927-86c9-4f3711183e1c/volumes" Nov 24 21:04:57 crc kubenswrapper[5035]: I1124 21:04:57.200092 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:04:57 crc kubenswrapper[5035]: E1124 21:04:57.200783 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:05:10 crc kubenswrapper[5035]: I1124 21:05:10.200252 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:05:10 crc kubenswrapper[5035]: E1124 21:05:10.200948 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:05:21 crc kubenswrapper[5035]: I1124 21:05:21.200613 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:05:21 crc kubenswrapper[5035]: E1124 21:05:21.201380 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:05:34 crc kubenswrapper[5035]: I1124 21:05:34.202166 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:05:34 crc kubenswrapper[5035]: E1124 21:05:34.203006 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:05:45 crc kubenswrapper[5035]: I1124 21:05:45.200789 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:05:45 crc kubenswrapper[5035]: E1124 21:05:45.202259 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:05:58 crc kubenswrapper[5035]: I1124 21:05:58.200785 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:05:58 crc kubenswrapper[5035]: E1124 21:05:58.201441 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:06:09 crc kubenswrapper[5035]: I1124 21:06:09.199864 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:06:09 crc kubenswrapper[5035]: E1124 21:06:09.201412 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:06:22 crc kubenswrapper[5035]: I1124 21:06:22.221280 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:06:22 crc kubenswrapper[5035]: I1124 21:06:22.565167 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"0042e755c9009b589a5cea8cd858a3c912c03b9d8d07e4a569dfd3a470a586d0"} Nov 24 21:08:40 crc kubenswrapper[5035]: I1124 21:08:40.991252 5035 generic.go:334] "Generic (PLEG): container finished" podID="7af30223-20cb-4470-b630-170318d5de9e" containerID="1d61a498c3f890146ab9b4327269b4408b8c686db8561eceac4eb014353b691a" exitCode=1 Nov 24 21:08:40 crc kubenswrapper[5035]: I1124 21:08:40.991366 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"7af30223-20cb-4470-b630-170318d5de9e","Type":"ContainerDied","Data":"1d61a498c3f890146ab9b4327269b4408b8c686db8561eceac4eb014353b691a"} Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.486927 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.636641 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-public-key\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.636744 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-private-key\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.636779 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ceph\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.636888 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxprm\" (UniqueName: \"kubernetes.io/projected/7af30223-20cb-4470-b630-170318d5de9e-kube-api-access-xxprm\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.636910 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-test-operator-clouds-config\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.636931 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ca-certs\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.636960 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-config\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.637011 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-kubeconfig\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.637038 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.637063 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-openstack-config-secret\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.637174 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-temporary\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.637213 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-workdir\") pod \"7af30223-20cb-4470-b630-170318d5de9e\" (UID: \"7af30223-20cb-4470-b630-170318d5de9e\") " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.638790 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.642398 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af30223-20cb-4470-b630-170318d5de9e-kube-api-access-xxprm" (OuterVolumeSpecName: "kube-api-access-xxprm") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "kube-api-access-xxprm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.644514 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ceph" (OuterVolumeSpecName: "ceph") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.645914 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.671941 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.718792 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.720766 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.727464 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.735506 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739385 5035 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739417 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739432 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxprm\" (UniqueName: \"kubernetes.io/projected/7af30223-20cb-4470-b630-170318d5de9e-kube-api-access-xxprm\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739444 5035 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-config\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739456 5035 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-kubeconfig\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739489 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739502 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739514 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.739527 5035 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.767399 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.779035 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.794575 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.840770 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/7af30223-20cb-4470-b630-170318d5de9e-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.840796 5035 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7af30223-20cb-4470-b630-170318d5de9e-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:42 crc kubenswrapper[5035]: I1124 21:08:42.840806 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:43 crc kubenswrapper[5035]: I1124 21:08:43.016508 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"7af30223-20cb-4470-b630-170318d5de9e","Type":"ContainerDied","Data":"8ebbe115cb8b60416502ccdc02a01c49fdc11b7e69a9089d77b27546fe8e9d6c"} Nov 24 21:08:43 crc kubenswrapper[5035]: I1124 21:08:43.016554 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ebbe115cb8b60416502ccdc02a01c49fdc11b7e69a9089d77b27546fe8e9d6c" Nov 24 21:08:43 crc kubenswrapper[5035]: I1124 21:08:43.017074 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Nov 24 21:08:43 crc kubenswrapper[5035]: I1124 21:08:43.941758 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "7af30223-20cb-4470-b630-170318d5de9e" (UID: "7af30223-20cb-4470-b630-170318d5de9e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:08:43 crc kubenswrapper[5035]: I1124 21:08:43.972543 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7af30223-20cb-4470-b630-170318d5de9e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 21:08:45 crc kubenswrapper[5035]: I1124 21:08:45.234267 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:08:45 crc kubenswrapper[5035]: I1124 21:08:45.234795 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.368166 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Nov 24 21:08:52 crc kubenswrapper[5035]: E1124 21:08:52.369527 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="extract-utilities" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.369551 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="extract-utilities" Nov 24 21:08:52 crc kubenswrapper[5035]: E1124 21:08:52.369585 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="registry-server" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.369598 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="registry-server" Nov 24 21:08:52 crc kubenswrapper[5035]: E1124 21:08:52.369619 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af30223-20cb-4470-b630-170318d5de9e" containerName="tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.369633 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af30223-20cb-4470-b630-170318d5de9e" containerName="tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: E1124 21:08:52.369655 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="extract-content" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.369666 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="extract-content" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.369972 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af30223-20cb-4470-b630-170318d5de9e" containerName="tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.370007 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc5984b-4d6f-4927-86c9-4f3711183e1c" containerName="registry-server" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.371168 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.383678 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.443893 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q658j\" (UniqueName: \"kubernetes.io/projected/b2081aca-59eb-43e9-b9fe-753837e7e553-kube-api-access-q658j\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"b2081aca-59eb-43e9-b9fe-753837e7e553\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.444011 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"b2081aca-59eb-43e9-b9fe-753837e7e553\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.546350 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q658j\" (UniqueName: \"kubernetes.io/projected/b2081aca-59eb-43e9-b9fe-753837e7e553-kube-api-access-q658j\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"b2081aca-59eb-43e9-b9fe-753837e7e553\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.546515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"b2081aca-59eb-43e9-b9fe-753837e7e553\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.547339 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"b2081aca-59eb-43e9-b9fe-753837e7e553\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.796358 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"b2081aca-59eb-43e9-b9fe-753837e7e553\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:52 crc kubenswrapper[5035]: I1124 21:08:52.798028 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q658j\" (UniqueName: \"kubernetes.io/projected/b2081aca-59eb-43e9-b9fe-753837e7e553-kube-api-access-q658j\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"b2081aca-59eb-43e9-b9fe-753837e7e553\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:53 crc kubenswrapper[5035]: I1124 21:08:53.088965 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Nov 24 21:08:53 crc kubenswrapper[5035]: I1124 21:08:53.643022 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Nov 24 21:08:54 crc kubenswrapper[5035]: I1124 21:08:54.144166 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"b2081aca-59eb-43e9-b9fe-753837e7e553","Type":"ContainerStarted","Data":"4687c8b5d80cac98d2ef181c5b182a69721d9e714e0b2c023144aed97a34084a"} Nov 24 21:08:55 crc kubenswrapper[5035]: I1124 21:08:55.169753 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"b2081aca-59eb-43e9-b9fe-753837e7e553","Type":"ContainerStarted","Data":"521906bda5f158199ad3b05e98a8bd083c7da674fd9255e810b30084f93eb1cf"} Nov 24 21:08:55 crc kubenswrapper[5035]: I1124 21:08:55.203410 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" podStartSLOduration=2.685492197 podStartE2EDuration="3.203380558s" podCreationTimestamp="2025-11-24 21:08:52 +0000 UTC" firstStartedPulling="2025-11-24 21:08:53.645166312 +0000 UTC m=+6872.167672609" lastFinishedPulling="2025-11-24 21:08:54.163054673 +0000 UTC m=+6872.685560970" observedRunningTime="2025-11-24 21:08:55.192646482 +0000 UTC m=+6873.715152759" watchObservedRunningTime="2025-11-24 21:08:55.203380558 +0000 UTC m=+6873.725886855" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.032969 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ansibletest-ansibletest"] Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.036079 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.036327 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.062226 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.062357 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.062850 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.063107 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164657 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164703 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164720 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164744 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164780 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164818 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164844 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164867 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164918 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs52v\" (UniqueName: \"kubernetes.io/projected/0eab5718-a1b7-4621-96bd-b35f01b4ed48-kube-api-access-qs52v\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.164950 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.166175 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.173111 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs52v\" (UniqueName: \"kubernetes.io/projected/0eab5718-a1b7-4621-96bd-b35f01b4ed48-kube-api-access-qs52v\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266414 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266482 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266506 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266533 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266559 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266639 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.266666 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.267375 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.267514 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.267637 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.271498 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.272099 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.272553 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.277278 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.285078 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs52v\" (UniqueName: \"kubernetes.io/projected/0eab5718-a1b7-4621-96bd-b35f01b4ed48-kube-api-access-qs52v\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.338790 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ansibletest-ansibletest\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.404672 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.886898 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Nov 24 21:09:07 crc kubenswrapper[5035]: I1124 21:09:07.898618 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 21:09:08 crc kubenswrapper[5035]: I1124 21:09:08.312209 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0eab5718-a1b7-4621-96bd-b35f01b4ed48","Type":"ContainerStarted","Data":"62f139084e7bfabf3e23d0acb356e3f20df80cbf749dee06432d135e567b75d9"} Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.393003 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gz6tz"] Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.400448 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.414443 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gz6tz"] Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.577025 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-catalog-content\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.577131 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-utilities\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.577206 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79dqk\" (UniqueName: \"kubernetes.io/projected/7f09a28f-95cb-4547-aab6-82589d6df0d2-kube-api-access-79dqk\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.679064 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-catalog-content\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.679163 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-utilities\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.679245 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79dqk\" (UniqueName: \"kubernetes.io/projected/7f09a28f-95cb-4547-aab6-82589d6df0d2-kube-api-access-79dqk\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.679679 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-catalog-content\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.679707 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-utilities\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.720821 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79dqk\" (UniqueName: \"kubernetes.io/projected/7f09a28f-95cb-4547-aab6-82589d6df0d2-kube-api-access-79dqk\") pod \"community-operators-gz6tz\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:12 crc kubenswrapper[5035]: I1124 21:09:12.727255 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:13 crc kubenswrapper[5035]: W1124 21:09:13.284635 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f09a28f_95cb_4547_aab6_82589d6df0d2.slice/crio-ead05b9097b05cede1e344f799de5082c0acf419da9ca2245d22ac67a206ee5f WatchSource:0}: Error finding container ead05b9097b05cede1e344f799de5082c0acf419da9ca2245d22ac67a206ee5f: Status 404 returned error can't find the container with id ead05b9097b05cede1e344f799de5082c0acf419da9ca2245d22ac67a206ee5f Nov 24 21:09:13 crc kubenswrapper[5035]: I1124 21:09:13.285457 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gz6tz"] Nov 24 21:09:13 crc kubenswrapper[5035]: I1124 21:09:13.389356 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz6tz" event={"ID":"7f09a28f-95cb-4547-aab6-82589d6df0d2","Type":"ContainerStarted","Data":"ead05b9097b05cede1e344f799de5082c0acf419da9ca2245d22ac67a206ee5f"} Nov 24 21:09:14 crc kubenswrapper[5035]: I1124 21:09:14.399338 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerID="6488170f9b2b5fa2f3e6508b60991a3be71a918f3b1af1a7901c5878cb8eca10" exitCode=0 Nov 24 21:09:14 crc kubenswrapper[5035]: I1124 21:09:14.399443 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz6tz" event={"ID":"7f09a28f-95cb-4547-aab6-82589d6df0d2","Type":"ContainerDied","Data":"6488170f9b2b5fa2f3e6508b60991a3be71a918f3b1af1a7901c5878cb8eca10"} Nov 24 21:09:15 crc kubenswrapper[5035]: I1124 21:09:15.234462 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:09:15 crc kubenswrapper[5035]: I1124 21:09:15.234523 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:09:15 crc kubenswrapper[5035]: I1124 21:09:15.411371 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz6tz" event={"ID":"7f09a28f-95cb-4547-aab6-82589d6df0d2","Type":"ContainerStarted","Data":"94436213419e1c4ab607449388acc284531615861bb9aca2b5bdd0a45e92f15f"} Nov 24 21:09:16 crc kubenswrapper[5035]: I1124 21:09:16.424494 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerID="94436213419e1c4ab607449388acc284531615861bb9aca2b5bdd0a45e92f15f" exitCode=0 Nov 24 21:09:16 crc kubenswrapper[5035]: I1124 21:09:16.424553 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz6tz" event={"ID":"7f09a28f-95cb-4547-aab6-82589d6df0d2","Type":"ContainerDied","Data":"94436213419e1c4ab607449388acc284531615861bb9aca2b5bdd0a45e92f15f"} Nov 24 21:09:17 crc kubenswrapper[5035]: I1124 21:09:17.445464 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz6tz" event={"ID":"7f09a28f-95cb-4547-aab6-82589d6df0d2","Type":"ContainerStarted","Data":"56f34ad8b0bfab5f983d0d58d8dfbf7ee5749242bd504950abe6fa0280891a70"} Nov 24 21:09:17 crc kubenswrapper[5035]: I1124 21:09:17.479865 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gz6tz" podStartSLOduration=3.083524753 podStartE2EDuration="5.47984425s" podCreationTimestamp="2025-11-24 21:09:12 +0000 UTC" firstStartedPulling="2025-11-24 21:09:14.401371021 +0000 UTC m=+6892.923877278" lastFinishedPulling="2025-11-24 21:09:16.797690498 +0000 UTC m=+6895.320196775" observedRunningTime="2025-11-24 21:09:17.466927246 +0000 UTC m=+6895.989433503" watchObservedRunningTime="2025-11-24 21:09:17.47984425 +0000 UTC m=+6896.002350507" Nov 24 21:09:22 crc kubenswrapper[5035]: I1124 21:09:22.727781 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:22 crc kubenswrapper[5035]: I1124 21:09:22.728390 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:22 crc kubenswrapper[5035]: I1124 21:09:22.795259 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:23 crc kubenswrapper[5035]: I1124 21:09:23.549938 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:23 crc kubenswrapper[5035]: I1124 21:09:23.608011 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gz6tz"] Nov 24 21:09:25 crc kubenswrapper[5035]: I1124 21:09:25.522279 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gz6tz" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="registry-server" containerID="cri-o://56f34ad8b0bfab5f983d0d58d8dfbf7ee5749242bd504950abe6fa0280891a70" gracePeriod=2 Nov 24 21:09:26 crc kubenswrapper[5035]: I1124 21:09:26.544976 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerID="56f34ad8b0bfab5f983d0d58d8dfbf7ee5749242bd504950abe6fa0280891a70" exitCode=0 Nov 24 21:09:26 crc kubenswrapper[5035]: I1124 21:09:26.545046 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz6tz" event={"ID":"7f09a28f-95cb-4547-aab6-82589d6df0d2","Type":"ContainerDied","Data":"56f34ad8b0bfab5f983d0d58d8dfbf7ee5749242bd504950abe6fa0280891a70"} Nov 24 21:09:27 crc kubenswrapper[5035]: E1124 21:09:27.671619 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified" Nov 24 21:09:27 crc kubenswrapper[5035]: E1124 21:09:27.672078 5035 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 24 21:09:27 crc kubenswrapper[5035]: container &Container{Name:ansibletest-ansibletest,Image:quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_ANSIBLE_EXTRA_VARS,Value:-e manual_run=false,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_FILE_EXTRA_VARS,Value:--- Nov 24 21:09:27 crc kubenswrapper[5035]: foo: bar Nov 24 21:09:27 crc kubenswrapper[5035]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_BRANCH,Value:,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_REPO,Value:https://github.com/ansible/test-playbooks,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_INVENTORY,Value:localhost ansible_connection=local ansible_python_interpreter=python3 Nov 24 21:09:27 crc kubenswrapper[5035]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_PLAYBOOK,Value:./debug.yml,ValueFrom:nil,},EnvVar{Name:POD_DEBUG,Value:false,ValueFrom:nil,},EnvVar{Name:POD_INSTALL_COLLECTIONS,Value:,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/ansible,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/AnsibleTests/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/ansible/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/var/lib/ansible/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:workload-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/test_keypair.key,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:compute-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/.ssh/compute_id,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qs52v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*227,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*227,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ansibletest-ansibletest_openstack(0eab5718-a1b7-4621-96bd-b35f01b4ed48): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 24 21:09:27 crc kubenswrapper[5035]: > logger="UnhandledError" Nov 24 21:09:27 crc kubenswrapper[5035]: E1124 21:09:27.673307 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ansibletest-ansibletest" podUID="0eab5718-a1b7-4621-96bd-b35f01b4ed48" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.002552 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.105330 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79dqk\" (UniqueName: \"kubernetes.io/projected/7f09a28f-95cb-4547-aab6-82589d6df0d2-kube-api-access-79dqk\") pod \"7f09a28f-95cb-4547-aab6-82589d6df0d2\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.105522 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-utilities\") pod \"7f09a28f-95cb-4547-aab6-82589d6df0d2\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.105844 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-catalog-content\") pod \"7f09a28f-95cb-4547-aab6-82589d6df0d2\" (UID: \"7f09a28f-95cb-4547-aab6-82589d6df0d2\") " Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.106470 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-utilities" (OuterVolumeSpecName: "utilities") pod "7f09a28f-95cb-4547-aab6-82589d6df0d2" (UID: "7f09a28f-95cb-4547-aab6-82589d6df0d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.138821 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f09a28f-95cb-4547-aab6-82589d6df0d2-kube-api-access-79dqk" (OuterVolumeSpecName: "kube-api-access-79dqk") pod "7f09a28f-95cb-4547-aab6-82589d6df0d2" (UID: "7f09a28f-95cb-4547-aab6-82589d6df0d2"). InnerVolumeSpecName "kube-api-access-79dqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.168244 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f09a28f-95cb-4547-aab6-82589d6df0d2" (UID: "7f09a28f-95cb-4547-aab6-82589d6df0d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.208404 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.208441 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79dqk\" (UniqueName: \"kubernetes.io/projected/7f09a28f-95cb-4547-aab6-82589d6df0d2-kube-api-access-79dqk\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.208456 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f09a28f-95cb-4547-aab6-82589d6df0d2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.564911 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gz6tz" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.565435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gz6tz" event={"ID":"7f09a28f-95cb-4547-aab6-82589d6df0d2","Type":"ContainerDied","Data":"ead05b9097b05cede1e344f799de5082c0acf419da9ca2245d22ac67a206ee5f"} Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.565475 5035 scope.go:117] "RemoveContainer" containerID="56f34ad8b0bfab5f983d0d58d8dfbf7ee5749242bd504950abe6fa0280891a70" Nov 24 21:09:28 crc kubenswrapper[5035]: E1124 21:09:28.566955 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified\\\"\"" pod="openstack/ansibletest-ansibletest" podUID="0eab5718-a1b7-4621-96bd-b35f01b4ed48" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.587190 5035 scope.go:117] "RemoveContainer" containerID="94436213419e1c4ab607449388acc284531615861bb9aca2b5bdd0a45e92f15f" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.625803 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gz6tz"] Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.652583 5035 scope.go:117] "RemoveContainer" containerID="6488170f9b2b5fa2f3e6508b60991a3be71a918f3b1af1a7901c5878cb8eca10" Nov 24 21:09:28 crc kubenswrapper[5035]: I1124 21:09:28.653081 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gz6tz"] Nov 24 21:09:30 crc kubenswrapper[5035]: I1124 21:09:30.211779 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" path="/var/lib/kubelet/pods/7f09a28f-95cb-4547-aab6-82589d6df0d2/volumes" Nov 24 21:09:42 crc kubenswrapper[5035]: I1124 21:09:42.766781 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0eab5718-a1b7-4621-96bd-b35f01b4ed48","Type":"ContainerStarted","Data":"e5d9a4974c612b77d9e440e819554f52fab965ac904ff8b89c06a9c1e969425c"} Nov 24 21:09:42 crc kubenswrapper[5035]: I1124 21:09:42.800033 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ansibletest-ansibletest" podStartSLOduration=4.047367553 podStartE2EDuration="37.800011905s" podCreationTimestamp="2025-11-24 21:09:05 +0000 UTC" firstStartedPulling="2025-11-24 21:09:07.898166086 +0000 UTC m=+6886.420672343" lastFinishedPulling="2025-11-24 21:09:41.650810398 +0000 UTC m=+6920.173316695" observedRunningTime="2025-11-24 21:09:42.789838794 +0000 UTC m=+6921.312345071" watchObservedRunningTime="2025-11-24 21:09:42.800011905 +0000 UTC m=+6921.322518162" Nov 24 21:09:44 crc kubenswrapper[5035]: I1124 21:09:44.801149 5035 generic.go:334] "Generic (PLEG): container finished" podID="0eab5718-a1b7-4621-96bd-b35f01b4ed48" containerID="e5d9a4974c612b77d9e440e819554f52fab965ac904ff8b89c06a9c1e969425c" exitCode=0 Nov 24 21:09:44 crc kubenswrapper[5035]: I1124 21:09:44.801247 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0eab5718-a1b7-4621-96bd-b35f01b4ed48","Type":"ContainerDied","Data":"e5d9a4974c612b77d9e440e819554f52fab965ac904ff8b89c06a9c1e969425c"} Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.234441 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.234847 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.235066 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.236501 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0042e755c9009b589a5cea8cd858a3c912c03b9d8d07e4a569dfd3a470a586d0"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.236780 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://0042e755c9009b589a5cea8cd858a3c912c03b9d8d07e4a569dfd3a470a586d0" gracePeriod=600 Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.821667 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="0042e755c9009b589a5cea8cd858a3c912c03b9d8d07e4a569dfd3a470a586d0" exitCode=0 Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.822223 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"0042e755c9009b589a5cea8cd858a3c912c03b9d8d07e4a569dfd3a470a586d0"} Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.822269 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e"} Nov 24 21:09:45 crc kubenswrapper[5035]: I1124 21:09:45.822323 5035 scope.go:117] "RemoveContainer" containerID="af68cf55796d897b283b5f605a12c21cd2f9138c36a5563d6ffe1361d94ef1b3" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.187159 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.195958 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196014 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196050 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-workload-ssh-secret\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196079 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-workdir\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196104 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config-secret\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196143 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-temporary\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196176 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ceph\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196197 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-compute-ssh-secret\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196221 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs52v\" (UniqueName: \"kubernetes.io/projected/0eab5718-a1b7-4621-96bd-b35f01b4ed48-kube-api-access-qs52v\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.196254 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.197332 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.210323 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.212964 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ceph" (OuterVolumeSpecName: "ceph") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.226224 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.232477 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eab5718-a1b7-4621-96bd-b35f01b4ed48-kube-api-access-qs52v" (OuterVolumeSpecName: "kube-api-access-qs52v") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "kube-api-access-qs52v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.239201 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.240418 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-workload-ssh-secret" (OuterVolumeSpecName: "workload-ssh-secret") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "workload-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.241802 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-compute-ssh-secret" (OuterVolumeSpecName: "compute-ssh-secret") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "compute-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: E1124 21:09:46.263189 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs podName:0eab5718-a1b7-4621-96bd-b35f01b4ed48 nodeName:}" failed. No retries permitted until 2025-11-24 21:09:46.763163885 +0000 UTC m=+6925.285670142 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ca-certs" (UniqueName: "kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48") : error deleting /var/lib/kubelet/pods/0eab5718-a1b7-4621-96bd-b35f01b4ed48/volume-subpaths: remove /var/lib/kubelet/pods/0eab5718-a1b7-4621-96bd-b35f01b4ed48/volume-subpaths: no such file or directory Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.263563 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298689 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298735 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298746 5035 reconciler_common.go:293] "Volume detached for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-workload-ssh-secret\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298756 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298766 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298780 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0eab5718-a1b7-4621-96bd-b35f01b4ed48-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298791 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298799 5035 reconciler_common.go:293] "Volume detached for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-compute-ssh-secret\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.298809 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs52v\" (UniqueName: \"kubernetes.io/projected/0eab5718-a1b7-4621-96bd-b35f01b4ed48-kube-api-access-qs52v\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.317835 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.401869 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.810762 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs\") pod \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\" (UID: \"0eab5718-a1b7-4621-96bd-b35f01b4ed48\") " Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.814619 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0eab5718-a1b7-4621-96bd-b35f01b4ed48" (UID: "0eab5718-a1b7-4621-96bd-b35f01b4ed48"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.835235 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0eab5718-a1b7-4621-96bd-b35f01b4ed48","Type":"ContainerDied","Data":"62f139084e7bfabf3e23d0acb356e3f20df80cbf749dee06432d135e567b75d9"} Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.835316 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62f139084e7bfabf3e23d0acb356e3f20df80cbf749dee06432d135e567b75d9" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.835346 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Nov 24 21:09:46 crc kubenswrapper[5035]: I1124 21:09:46.912870 5035 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0eab5718-a1b7-4621-96bd-b35f01b4ed48-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.855841 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Nov 24 21:09:52 crc kubenswrapper[5035]: E1124 21:09:52.857027 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="extract-content" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.857046 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="extract-content" Nov 24 21:09:52 crc kubenswrapper[5035]: E1124 21:09:52.857072 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="registry-server" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.857085 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="registry-server" Nov 24 21:09:52 crc kubenswrapper[5035]: E1124 21:09:52.857106 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="extract-utilities" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.857117 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="extract-utilities" Nov 24 21:09:52 crc kubenswrapper[5035]: E1124 21:09:52.857141 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eab5718-a1b7-4621-96bd-b35f01b4ed48" containerName="ansibletest-ansibletest" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.857148 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eab5718-a1b7-4621-96bd-b35f01b4ed48" containerName="ansibletest-ansibletest" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.857405 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f09a28f-95cb-4547-aab6-82589d6df0d2" containerName="registry-server" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.857435 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eab5718-a1b7-4621-96bd-b35f01b4ed48" containerName="ansibletest-ansibletest" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.858187 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:52 crc kubenswrapper[5035]: I1124 21:09:52.865331 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.026400 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"928c3f00-4054-4f1d-8693-102b6032a11a\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.026464 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzkcz\" (UniqueName: \"kubernetes.io/projected/928c3f00-4054-4f1d-8693-102b6032a11a-kube-api-access-qzkcz\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"928c3f00-4054-4f1d-8693-102b6032a11a\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.128168 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"928c3f00-4054-4f1d-8693-102b6032a11a\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.128451 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzkcz\" (UniqueName: \"kubernetes.io/projected/928c3f00-4054-4f1d-8693-102b6032a11a-kube-api-access-qzkcz\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"928c3f00-4054-4f1d-8693-102b6032a11a\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.128522 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"928c3f00-4054-4f1d-8693-102b6032a11a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.146887 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzkcz\" (UniqueName: \"kubernetes.io/projected/928c3f00-4054-4f1d-8693-102b6032a11a-kube-api-access-qzkcz\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"928c3f00-4054-4f1d-8693-102b6032a11a\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.152450 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"928c3f00-4054-4f1d-8693-102b6032a11a\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.184375 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.653282 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Nov 24 21:09:53 crc kubenswrapper[5035]: I1124 21:09:53.918264 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"928c3f00-4054-4f1d-8693-102b6032a11a","Type":"ContainerStarted","Data":"a4b639b0e2a4231caeef9a0135ae68edf820c324f3a66d3f8cb577274620150d"} Nov 24 21:09:54 crc kubenswrapper[5035]: I1124 21:09:54.932511 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"928c3f00-4054-4f1d-8693-102b6032a11a","Type":"ContainerStarted","Data":"f2fbb8c9ae6006b18786f61a1a9d004a410cd5aee0ab822c11c60f42e34ba424"} Nov 24 21:09:54 crc kubenswrapper[5035]: I1124 21:09:54.962557 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" podStartSLOduration=2.496614313 podStartE2EDuration="2.962537519s" podCreationTimestamp="2025-11-24 21:09:52 +0000 UTC" firstStartedPulling="2025-11-24 21:09:53.656215341 +0000 UTC m=+6932.178721598" lastFinishedPulling="2025-11-24 21:09:54.122138507 +0000 UTC m=+6932.644644804" observedRunningTime="2025-11-24 21:09:54.955530812 +0000 UTC m=+6933.478037109" watchObservedRunningTime="2025-11-24 21:09:54.962537519 +0000 UTC m=+6933.485043776" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.880637 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizontest-tests-horizontest"] Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.884435 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.887128 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizontest-tests-horizontesthorizontest-config" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.887248 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.894276 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961229 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961313 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961350 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z9qx\" (UniqueName: \"kubernetes.io/projected/36ff01ed-eaf8-488f-9243-44710cb7017f-kube-api-access-9z9qx\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961378 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961435 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961483 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961512 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:07 crc kubenswrapper[5035]: I1124 21:10:07.961547 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.081778 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.081850 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z9qx\" (UniqueName: \"kubernetes.io/projected/36ff01ed-eaf8-488f-9243-44710cb7017f-kube-api-access-9z9qx\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.081896 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.081981 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.082026 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.082071 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.082102 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.082230 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.082765 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.086022 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.090211 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.090961 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.093274 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.103920 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.104564 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.113037 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z9qx\" (UniqueName: \"kubernetes.io/projected/36ff01ed-eaf8-488f-9243-44710cb7017f-kube-api-access-9z9qx\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.178751 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.224344 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 24 21:10:08 crc kubenswrapper[5035]: I1124 21:10:08.677645 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Nov 24 21:10:09 crc kubenswrapper[5035]: I1124 21:10:09.102158 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"36ff01ed-eaf8-488f-9243-44710cb7017f","Type":"ContainerStarted","Data":"dc7b07adf0177ede52c3bc1d7509851ea69e6728bc18aa879e96194fc9f06644"} Nov 24 21:10:35 crc kubenswrapper[5035]: E1124 21:10:35.845418 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizontest:current-podified" Nov 24 21:10:35 crc kubenswrapper[5035]: E1124 21:10:35.847134 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizontest-tests-horizontest,Image:quay.io/podified-antelope-centos9/openstack-horizontest:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADMIN_PASSWORD,Value:12345678,ValueFrom:nil,},EnvVar{Name:ADMIN_USERNAME,Value:admin,ValueFrom:nil,},EnvVar{Name:AUTH_URL,Value:https://keystone-public-openstack.apps-crc.testing,ValueFrom:nil,},EnvVar{Name:DASHBOARD_URL,Value:https://horizon-openstack.apps-crc.testing/,ValueFrom:nil,},EnvVar{Name:EXTRA_FLAG,Value:not pagination and test_users.py,ValueFrom:nil,},EnvVar{Name:FLAVOR_NAME,Value:m1.tiny,ValueFrom:nil,},EnvVar{Name:HORIZONTEST_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:HORIZON_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:HORIZON_LOGS_DIR_NAME,Value:horizon,ValueFrom:nil,},EnvVar{Name:HORIZON_REPO_BRANCH,Value:master,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE,Value:/var/lib/horizontest/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE_NAME,Value:cirros-0.6.2-x86_64-disk,ValueFrom:nil,},EnvVar{Name:IMAGE_URL,Value:http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:PASSWORD,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME_XPATH,Value://*[@class=\"context-project\"]//ancestor::ul,ValueFrom:nil,},EnvVar{Name:REPO_URL,Value:https://review.opendev.org/openstack/horizon,ValueFrom:nil,},EnvVar{Name:USER_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/horizontest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/horizontest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/horizontest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9z9qx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42455,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42455,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizontest-tests-horizontest_openstack(36ff01ed-eaf8-488f-9243-44710cb7017f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 21:10:35 crc kubenswrapper[5035]: E1124 21:10:35.848411 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizontest-tests-horizontest" podUID="36ff01ed-eaf8-488f-9243-44710cb7017f" Nov 24 21:10:36 crc kubenswrapper[5035]: E1124 21:10:36.370771 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizontest:current-podified\\\"\"" pod="openstack/horizontest-tests-horizontest" podUID="36ff01ed-eaf8-488f-9243-44710cb7017f" Nov 24 21:10:52 crc kubenswrapper[5035]: I1124 21:10:52.538262 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"36ff01ed-eaf8-488f-9243-44710cb7017f","Type":"ContainerStarted","Data":"d86e9bc3cd32b7ce30d83ab76dad6b5f078095557134db5df9f08ef0ea47b4e5"} Nov 24 21:10:52 crc kubenswrapper[5035]: I1124 21:10:52.575749 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizontest-tests-horizontest" podStartSLOduration=4.485501983 podStartE2EDuration="46.575727245s" podCreationTimestamp="2025-11-24 21:10:06 +0000 UTC" firstStartedPulling="2025-11-24 21:10:08.684976754 +0000 UTC m=+6947.207483031" lastFinishedPulling="2025-11-24 21:10:50.775202046 +0000 UTC m=+6989.297708293" observedRunningTime="2025-11-24 21:10:52.568372349 +0000 UTC m=+6991.090878616" watchObservedRunningTime="2025-11-24 21:10:52.575727245 +0000 UTC m=+6991.098233512" Nov 24 21:11:45 crc kubenswrapper[5035]: I1124 21:11:45.234177 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:11:45 crc kubenswrapper[5035]: I1124 21:11:45.234705 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:12:15 crc kubenswrapper[5035]: I1124 21:12:15.234718 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:12:15 crc kubenswrapper[5035]: I1124 21:12:15.235401 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.248556 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.249211 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.249267 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.250520 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.250609 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" gracePeriod=600 Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.596710 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" exitCode=0 Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.596759 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e"} Nov 24 21:12:45 crc kubenswrapper[5035]: I1124 21:12:45.596795 5035 scope.go:117] "RemoveContainer" containerID="0042e755c9009b589a5cea8cd858a3c912c03b9d8d07e4a569dfd3a470a586d0" Nov 24 21:12:45 crc kubenswrapper[5035]: E1124 21:12:45.957020 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:12:46 crc kubenswrapper[5035]: I1124 21:12:46.606633 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:12:46 crc kubenswrapper[5035]: E1124 21:12:46.606903 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:12:50 crc kubenswrapper[5035]: I1124 21:12:50.638458 5035 generic.go:334] "Generic (PLEG): container finished" podID="36ff01ed-eaf8-488f-9243-44710cb7017f" containerID="d86e9bc3cd32b7ce30d83ab76dad6b5f078095557134db5df9f08ef0ea47b4e5" exitCode=0 Nov 24 21:12:50 crc kubenswrapper[5035]: I1124 21:12:50.638602 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"36ff01ed-eaf8-488f-9243-44710cb7017f","Type":"ContainerDied","Data":"d86e9bc3cd32b7ce30d83ab76dad6b5f078095557134db5df9f08ef0ea47b4e5"} Nov 24 21:12:51 crc kubenswrapper[5035]: I1124 21:12:51.996008 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.141429 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.141833 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ca-certs\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.142071 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-clouds-config\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.142459 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ceph\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.142707 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z9qx\" (UniqueName: \"kubernetes.io/projected/36ff01ed-eaf8-488f-9243-44710cb7017f-kube-api-access-9z9qx\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.143046 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-temporary\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.143604 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-workdir\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.143918 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-openstack-config-secret\") pod \"36ff01ed-eaf8-488f-9243-44710cb7017f\" (UID: \"36ff01ed-eaf8-488f-9243-44710cb7017f\") " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.144016 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.145564 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.161365 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36ff01ed-eaf8-488f-9243-44710cb7017f-kube-api-access-9z9qx" (OuterVolumeSpecName: "kube-api-access-9z9qx") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "kube-api-access-9z9qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.162232 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.163021 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ceph" (OuterVolumeSpecName: "ceph") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.190556 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.215460 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.222793 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.247900 5035 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.248190 5035 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.248224 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.248238 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.248279 5035 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/36ff01ed-eaf8-488f-9243-44710cb7017f-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.248306 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z9qx\" (UniqueName: \"kubernetes.io/projected/36ff01ed-eaf8-488f-9243-44710cb7017f-kube-api-access-9z9qx\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.268133 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.350970 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.408443 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "36ff01ed-eaf8-488f-9243-44710cb7017f" (UID: "36ff01ed-eaf8-488f-9243-44710cb7017f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.452809 5035 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/36ff01ed-eaf8-488f-9243-44710cb7017f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.657375 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.657364 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"36ff01ed-eaf8-488f-9243-44710cb7017f","Type":"ContainerDied","Data":"dc7b07adf0177ede52c3bc1d7509851ea69e6728bc18aa879e96194fc9f06644"} Nov 24 21:12:52 crc kubenswrapper[5035]: I1124 21:12:52.657500 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc7b07adf0177ede52c3bc1d7509851ea69e6728bc18aa879e96194fc9f06644" Nov 24 21:13:01 crc kubenswrapper[5035]: I1124 21:13:01.200609 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:13:01 crc kubenswrapper[5035]: E1124 21:13:01.201691 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.343747 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Nov 24 21:13:02 crc kubenswrapper[5035]: E1124 21:13:02.344915 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36ff01ed-eaf8-488f-9243-44710cb7017f" containerName="horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.345039 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="36ff01ed-eaf8-488f-9243-44710cb7017f" containerName="horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.345437 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="36ff01ed-eaf8-488f-9243-44710cb7017f" containerName="horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.346402 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.354749 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.488998 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqdpp\" (UniqueName: \"kubernetes.io/projected/c50bbf30-2132-45d9-93fb-4d266d85bf00-kube-api-access-gqdpp\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"c50bbf30-2132-45d9-93fb-4d266d85bf00\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.489537 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"c50bbf30-2132-45d9-93fb-4d266d85bf00\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.592187 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"c50bbf30-2132-45d9-93fb-4d266d85bf00\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.592789 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqdpp\" (UniqueName: \"kubernetes.io/projected/c50bbf30-2132-45d9-93fb-4d266d85bf00-kube-api-access-gqdpp\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"c50bbf30-2132-45d9-93fb-4d266d85bf00\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.592979 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"c50bbf30-2132-45d9-93fb-4d266d85bf00\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.634630 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqdpp\" (UniqueName: \"kubernetes.io/projected/c50bbf30-2132-45d9-93fb-4d266d85bf00-kube-api-access-gqdpp\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"c50bbf30-2132-45d9-93fb-4d266d85bf00\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.639610 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"c50bbf30-2132-45d9-93fb-4d266d85bf00\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: I1124 21:13:02.715836 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Nov 24 21:13:02 crc kubenswrapper[5035]: E1124 21:13:02.715930 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:13:03 crc kubenswrapper[5035]: I1124 21:13:03.207672 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Nov 24 21:13:03 crc kubenswrapper[5035]: E1124 21:13:03.214854 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:13:03 crc kubenswrapper[5035]: I1124 21:13:03.781258 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"c50bbf30-2132-45d9-93fb-4d266d85bf00","Type":"ContainerStarted","Data":"bb293161f1701108370e7e50f1398e857724e5ad512b34d25e725a7ec605804d"} Nov 24 21:13:04 crc kubenswrapper[5035]: E1124 21:13:04.372636 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:13:05 crc kubenswrapper[5035]: I1124 21:13:05.801631 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"c50bbf30-2132-45d9-93fb-4d266d85bf00","Type":"ContainerStarted","Data":"38c9689642dc3fa13104e8d0914744ca39a62acb0fe25d016538d0100cb0fbc7"} Nov 24 21:13:05 crc kubenswrapper[5035]: E1124 21:13:05.802788 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:13:05 crc kubenswrapper[5035]: I1124 21:13:05.822369 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" podStartSLOduration=2.666536294 podStartE2EDuration="3.822345419s" podCreationTimestamp="2025-11-24 21:13:02 +0000 UTC" firstStartedPulling="2025-11-24 21:13:03.216718559 +0000 UTC m=+7121.739224826" lastFinishedPulling="2025-11-24 21:13:04.372527654 +0000 UTC m=+7122.895033951" observedRunningTime="2025-11-24 21:13:05.821865977 +0000 UTC m=+7124.344372264" watchObservedRunningTime="2025-11-24 21:13:05.822345419 +0000 UTC m=+7124.344851686" Nov 24 21:13:06 crc kubenswrapper[5035]: E1124 21:13:06.813195 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:13:14 crc kubenswrapper[5035]: I1124 21:13:14.200635 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:13:14 crc kubenswrapper[5035]: E1124 21:13:14.201543 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:13:26 crc kubenswrapper[5035]: I1124 21:13:26.200474 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:13:26 crc kubenswrapper[5035]: E1124 21:13:26.201282 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:13:37 crc kubenswrapper[5035]: I1124 21:13:37.202206 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:13:37 crc kubenswrapper[5035]: E1124 21:13:37.203431 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.448467 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pw2tc/must-gather-l46g5"] Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.451162 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.452532 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pw2tc"/"kube-root-ca.crt" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.452963 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pw2tc"/"openshift-service-ca.crt" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.453227 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pw2tc"/"default-dockercfg-x5cxf" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.458483 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pw2tc/must-gather-l46g5"] Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.536551 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b1c38fb-066c-4261-8e53-a38f6e3ac622-must-gather-output\") pod \"must-gather-l46g5\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.536698 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pfkk\" (UniqueName: \"kubernetes.io/projected/6b1c38fb-066c-4261-8e53-a38f6e3ac622-kube-api-access-6pfkk\") pod \"must-gather-l46g5\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.638750 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pfkk\" (UniqueName: \"kubernetes.io/projected/6b1c38fb-066c-4261-8e53-a38f6e3ac622-kube-api-access-6pfkk\") pod \"must-gather-l46g5\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.639160 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b1c38fb-066c-4261-8e53-a38f6e3ac622-must-gather-output\") pod \"must-gather-l46g5\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.639530 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b1c38fb-066c-4261-8e53-a38f6e3ac622-must-gather-output\") pod \"must-gather-l46g5\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.660529 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pfkk\" (UniqueName: \"kubernetes.io/projected/6b1c38fb-066c-4261-8e53-a38f6e3ac622-kube-api-access-6pfkk\") pod \"must-gather-l46g5\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:45 crc kubenswrapper[5035]: I1124 21:13:45.769654 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:13:46 crc kubenswrapper[5035]: I1124 21:13:46.236773 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pw2tc/must-gather-l46g5"] Nov 24 21:13:47 crc kubenswrapper[5035]: I1124 21:13:47.208922 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/must-gather-l46g5" event={"ID":"6b1c38fb-066c-4261-8e53-a38f6e3ac622","Type":"ContainerStarted","Data":"cd10d5f86c78fbfb2c9d2012607b00beb44f8e1d2a721410aa0114a8b9f5eeb6"} Nov 24 21:13:51 crc kubenswrapper[5035]: I1124 21:13:51.248388 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/must-gather-l46g5" event={"ID":"6b1c38fb-066c-4261-8e53-a38f6e3ac622","Type":"ContainerStarted","Data":"9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982"} Nov 24 21:13:51 crc kubenswrapper[5035]: I1124 21:13:51.248993 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/must-gather-l46g5" event={"ID":"6b1c38fb-066c-4261-8e53-a38f6e3ac622","Type":"ContainerStarted","Data":"617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac"} Nov 24 21:13:51 crc kubenswrapper[5035]: I1124 21:13:51.271181 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pw2tc/must-gather-l46g5" podStartSLOduration=2.25771055 podStartE2EDuration="6.271161706s" podCreationTimestamp="2025-11-24 21:13:45 +0000 UTC" firstStartedPulling="2025-11-24 21:13:46.241492988 +0000 UTC m=+7164.763999245" lastFinishedPulling="2025-11-24 21:13:50.254944104 +0000 UTC m=+7168.777450401" observedRunningTime="2025-11-24 21:13:51.26235613 +0000 UTC m=+7169.784862387" watchObservedRunningTime="2025-11-24 21:13:51.271161706 +0000 UTC m=+7169.793667963" Nov 24 21:13:52 crc kubenswrapper[5035]: I1124 21:13:52.200318 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:13:52 crc kubenswrapper[5035]: E1124 21:13:52.200607 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.585212 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-b95zr"] Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.587482 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.654344 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-host\") pod \"crc-debug-b95zr\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.654824 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxgk7\" (UniqueName: \"kubernetes.io/projected/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-kube-api-access-kxgk7\") pod \"crc-debug-b95zr\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.756888 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-host\") pod \"crc-debug-b95zr\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.757025 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-host\") pod \"crc-debug-b95zr\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.757109 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxgk7\" (UniqueName: \"kubernetes.io/projected/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-kube-api-access-kxgk7\") pod \"crc-debug-b95zr\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.776639 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxgk7\" (UniqueName: \"kubernetes.io/projected/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-kube-api-access-kxgk7\") pod \"crc-debug-b95zr\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: I1124 21:13:55.908676 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:13:55 crc kubenswrapper[5035]: W1124 21:13:55.942452 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode786b372_cf0a_4004_b6a3_0a13e6fc66e8.slice/crio-0e30f9b0adce6cffe411d182254e7b56be63a7a29573c17ce4937e5461287722 WatchSource:0}: Error finding container 0e30f9b0adce6cffe411d182254e7b56be63a7a29573c17ce4937e5461287722: Status 404 returned error can't find the container with id 0e30f9b0adce6cffe411d182254e7b56be63a7a29573c17ce4937e5461287722 Nov 24 21:13:56 crc kubenswrapper[5035]: I1124 21:13:56.294423 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" event={"ID":"e786b372-cf0a-4004-b6a3-0a13e6fc66e8","Type":"ContainerStarted","Data":"0e30f9b0adce6cffe411d182254e7b56be63a7a29573c17ce4937e5461287722"} Nov 24 21:14:06 crc kubenswrapper[5035]: I1124 21:14:06.199936 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:14:06 crc kubenswrapper[5035]: E1124 21:14:06.200770 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:14:10 crc kubenswrapper[5035]: I1124 21:14:10.433533 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" event={"ID":"e786b372-cf0a-4004-b6a3-0a13e6fc66e8","Type":"ContainerStarted","Data":"7224554783c82c0a2b63712d7a321740192e0bff29db812e4c752143b003871f"} Nov 24 21:14:10 crc kubenswrapper[5035]: I1124 21:14:10.446768 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" podStartSLOduration=1.896297822 podStartE2EDuration="15.446749939s" podCreationTimestamp="2025-11-24 21:13:55 +0000 UTC" firstStartedPulling="2025-11-24 21:13:55.944867478 +0000 UTC m=+7174.467373745" lastFinishedPulling="2025-11-24 21:14:09.495319605 +0000 UTC m=+7188.017825862" observedRunningTime="2025-11-24 21:14:10.445036384 +0000 UTC m=+7188.967542641" watchObservedRunningTime="2025-11-24 21:14:10.446749939 +0000 UTC m=+7188.969256196" Nov 24 21:14:17 crc kubenswrapper[5035]: I1124 21:14:17.199692 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:14:17 crc kubenswrapper[5035]: E1124 21:14:17.200823 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:14:19 crc kubenswrapper[5035]: E1124 21:14:19.200823 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.085968 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8khr"] Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.088947 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.096899 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8khr"] Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.226892 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-utilities\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.227337 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnk8w\" (UniqueName: \"kubernetes.io/projected/80a3e8cf-9d55-48b6-9098-d4541e26d46b-kube-api-access-nnk8w\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.227362 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-catalog-content\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.329902 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnk8w\" (UniqueName: \"kubernetes.io/projected/80a3e8cf-9d55-48b6-9098-d4541e26d46b-kube-api-access-nnk8w\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.329998 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-catalog-content\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.330448 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-utilities\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.330737 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-catalog-content\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.330803 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-utilities\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.350846 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnk8w\" (UniqueName: \"kubernetes.io/projected/80a3e8cf-9d55-48b6-9098-d4541e26d46b-kube-api-access-nnk8w\") pod \"redhat-operators-g8khr\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:23 crc kubenswrapper[5035]: I1124 21:14:23.425768 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:24 crc kubenswrapper[5035]: I1124 21:14:24.018283 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8khr"] Nov 24 21:14:24 crc kubenswrapper[5035]: I1124 21:14:24.554956 5035 generic.go:334] "Generic (PLEG): container finished" podID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerID="8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd" exitCode=0 Nov 24 21:14:24 crc kubenswrapper[5035]: I1124 21:14:24.555187 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8khr" event={"ID":"80a3e8cf-9d55-48b6-9098-d4541e26d46b","Type":"ContainerDied","Data":"8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd"} Nov 24 21:14:24 crc kubenswrapper[5035]: I1124 21:14:24.555330 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8khr" event={"ID":"80a3e8cf-9d55-48b6-9098-d4541e26d46b","Type":"ContainerStarted","Data":"69ce6ff714596c0d3e9af45c388e8bd8d232fef8a9894df9b01cd4a6d0d28961"} Nov 24 21:14:24 crc kubenswrapper[5035]: I1124 21:14:24.569479 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 21:14:25 crc kubenswrapper[5035]: I1124 21:14:25.565848 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8khr" event={"ID":"80a3e8cf-9d55-48b6-9098-d4541e26d46b","Type":"ContainerStarted","Data":"3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae"} Nov 24 21:14:28 crc kubenswrapper[5035]: I1124 21:14:28.200306 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:14:28 crc kubenswrapper[5035]: E1124 21:14:28.201116 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:14:30 crc kubenswrapper[5035]: I1124 21:14:30.612709 5035 generic.go:334] "Generic (PLEG): container finished" podID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerID="3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae" exitCode=0 Nov 24 21:14:30 crc kubenswrapper[5035]: I1124 21:14:30.612807 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8khr" event={"ID":"80a3e8cf-9d55-48b6-9098-d4541e26d46b","Type":"ContainerDied","Data":"3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae"} Nov 24 21:14:31 crc kubenswrapper[5035]: I1124 21:14:31.622238 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8khr" event={"ID":"80a3e8cf-9d55-48b6-9098-d4541e26d46b","Type":"ContainerStarted","Data":"ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264"} Nov 24 21:14:31 crc kubenswrapper[5035]: I1124 21:14:31.644900 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8khr" podStartSLOduration=2.170257311 podStartE2EDuration="8.644875854s" podCreationTimestamp="2025-11-24 21:14:23 +0000 UTC" firstStartedPulling="2025-11-24 21:14:24.569150241 +0000 UTC m=+7203.091656498" lastFinishedPulling="2025-11-24 21:14:31.043768784 +0000 UTC m=+7209.566275041" observedRunningTime="2025-11-24 21:14:31.640014694 +0000 UTC m=+7210.162520951" watchObservedRunningTime="2025-11-24 21:14:31.644875854 +0000 UTC m=+7210.167382111" Nov 24 21:14:33 crc kubenswrapper[5035]: I1124 21:14:33.426703 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:33 crc kubenswrapper[5035]: I1124 21:14:33.427073 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:34 crc kubenswrapper[5035]: I1124 21:14:34.479001 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g8khr" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="registry-server" probeResult="failure" output=< Nov 24 21:14:34 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Nov 24 21:14:34 crc kubenswrapper[5035]: > Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.696460 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qzl4h"] Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.702531 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.706237 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzl4h"] Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.880882 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn4r5\" (UniqueName: \"kubernetes.io/projected/d27db1a1-f02b-4f19-a483-5aa0c79266dc-kube-api-access-sn4r5\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.881038 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-utilities\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.881167 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-catalog-content\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.982947 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn4r5\" (UniqueName: \"kubernetes.io/projected/d27db1a1-f02b-4f19-a483-5aa0c79266dc-kube-api-access-sn4r5\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.983074 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-utilities\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.983155 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-catalog-content\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.983736 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-utilities\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:39 crc kubenswrapper[5035]: I1124 21:14:39.983761 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-catalog-content\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:40 crc kubenswrapper[5035]: I1124 21:14:40.002244 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn4r5\" (UniqueName: \"kubernetes.io/projected/d27db1a1-f02b-4f19-a483-5aa0c79266dc-kube-api-access-sn4r5\") pod \"certified-operators-qzl4h\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:40 crc kubenswrapper[5035]: I1124 21:14:40.042235 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:40 crc kubenswrapper[5035]: I1124 21:14:40.522915 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzl4h"] Nov 24 21:14:40 crc kubenswrapper[5035]: I1124 21:14:40.718006 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerStarted","Data":"f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276"} Nov 24 21:14:40 crc kubenswrapper[5035]: I1124 21:14:40.718418 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerStarted","Data":"7455f04950a3e59a65010311097885948655ed6b1afdddc604e3809d2bc6465d"} Nov 24 21:14:41 crc kubenswrapper[5035]: I1124 21:14:41.727641 5035 generic.go:334] "Generic (PLEG): container finished" podID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerID="f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276" exitCode=0 Nov 24 21:14:41 crc kubenswrapper[5035]: I1124 21:14:41.727747 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerDied","Data":"f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276"} Nov 24 21:14:42 crc kubenswrapper[5035]: I1124 21:14:42.206170 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:14:42 crc kubenswrapper[5035]: E1124 21:14:42.206813 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:14:42 crc kubenswrapper[5035]: I1124 21:14:42.740515 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerStarted","Data":"33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5"} Nov 24 21:14:43 crc kubenswrapper[5035]: I1124 21:14:43.488334 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:43 crc kubenswrapper[5035]: I1124 21:14:43.536721 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.515732 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jzws2"] Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.517970 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.526432 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzws2"] Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.682445 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-catalog-content\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.682814 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldnzs\" (UniqueName: \"kubernetes.io/projected/02e6808c-5f1e-4a32-bea6-bec389db20c2-kube-api-access-ldnzs\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.682945 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-utilities\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.759183 5035 generic.go:334] "Generic (PLEG): container finished" podID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerID="33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5" exitCode=0 Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.759222 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerDied","Data":"33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5"} Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.784727 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-catalog-content\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.784812 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldnzs\" (UniqueName: \"kubernetes.io/projected/02e6808c-5f1e-4a32-bea6-bec389db20c2-kube-api-access-ldnzs\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.784945 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-utilities\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.785359 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-catalog-content\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.785493 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-utilities\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.822770 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldnzs\" (UniqueName: \"kubernetes.io/projected/02e6808c-5f1e-4a32-bea6-bec389db20c2-kube-api-access-ldnzs\") pod \"redhat-marketplace-jzws2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.855859 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.898710 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8khr"] Nov 24 21:14:44 crc kubenswrapper[5035]: I1124 21:14:44.898951 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8khr" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="registry-server" containerID="cri-o://ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264" gracePeriod=2 Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.396635 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:45 crc kubenswrapper[5035]: W1124 21:14:45.519375 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02e6808c_5f1e_4a32_bea6_bec389db20c2.slice/crio-39e4333863ea6806e72e11bb024349295dddd7825e4e5e9cc3e746d5dcdc818c WatchSource:0}: Error finding container 39e4333863ea6806e72e11bb024349295dddd7825e4e5e9cc3e746d5dcdc818c: Status 404 returned error can't find the container with id 39e4333863ea6806e72e11bb024349295dddd7825e4e5e9cc3e746d5dcdc818c Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.520426 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzws2"] Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.520439 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnk8w\" (UniqueName: \"kubernetes.io/projected/80a3e8cf-9d55-48b6-9098-d4541e26d46b-kube-api-access-nnk8w\") pod \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.520784 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-utilities\") pod \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.520854 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-catalog-content\") pod \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\" (UID: \"80a3e8cf-9d55-48b6-9098-d4541e26d46b\") " Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.521765 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-utilities" (OuterVolumeSpecName: "utilities") pod "80a3e8cf-9d55-48b6-9098-d4541e26d46b" (UID: "80a3e8cf-9d55-48b6-9098-d4541e26d46b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.527430 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80a3e8cf-9d55-48b6-9098-d4541e26d46b-kube-api-access-nnk8w" (OuterVolumeSpecName: "kube-api-access-nnk8w") pod "80a3e8cf-9d55-48b6-9098-d4541e26d46b" (UID: "80a3e8cf-9d55-48b6-9098-d4541e26d46b"). InnerVolumeSpecName "kube-api-access-nnk8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.530465 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnk8w\" (UniqueName: \"kubernetes.io/projected/80a3e8cf-9d55-48b6-9098-d4541e26d46b-kube-api-access-nnk8w\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.530506 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.627380 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80a3e8cf-9d55-48b6-9098-d4541e26d46b" (UID: "80a3e8cf-9d55-48b6-9098-d4541e26d46b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.632499 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80a3e8cf-9d55-48b6-9098-d4541e26d46b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.768842 5035 generic.go:334] "Generic (PLEG): container finished" podID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerID="ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264" exitCode=0 Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.768903 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8khr" event={"ID":"80a3e8cf-9d55-48b6-9098-d4541e26d46b","Type":"ContainerDied","Data":"ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264"} Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.768913 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8khr" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.768927 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8khr" event={"ID":"80a3e8cf-9d55-48b6-9098-d4541e26d46b","Type":"ContainerDied","Data":"69ce6ff714596c0d3e9af45c388e8bd8d232fef8a9894df9b01cd4a6d0d28961"} Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.768943 5035 scope.go:117] "RemoveContainer" containerID="ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.779253 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerStarted","Data":"ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6"} Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.781876 5035 generic.go:334] "Generic (PLEG): container finished" podID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerID="97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121" exitCode=0 Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.781917 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzws2" event={"ID":"02e6808c-5f1e-4a32-bea6-bec389db20c2","Type":"ContainerDied","Data":"97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121"} Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.781967 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzws2" event={"ID":"02e6808c-5f1e-4a32-bea6-bec389db20c2","Type":"ContainerStarted","Data":"39e4333863ea6806e72e11bb024349295dddd7825e4e5e9cc3e746d5dcdc818c"} Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.794654 5035 scope.go:117] "RemoveContainer" containerID="3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.813314 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qzl4h" podStartSLOduration=3.293949265 podStartE2EDuration="6.813271402s" podCreationTimestamp="2025-11-24 21:14:39 +0000 UTC" firstStartedPulling="2025-11-24 21:14:41.730098648 +0000 UTC m=+7220.252604905" lastFinishedPulling="2025-11-24 21:14:45.249420775 +0000 UTC m=+7223.771927042" observedRunningTime="2025-11-24 21:14:45.807140889 +0000 UTC m=+7224.329647156" watchObservedRunningTime="2025-11-24 21:14:45.813271402 +0000 UTC m=+7224.335777659" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.833987 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8khr"] Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.848835 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8khr"] Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.891196 5035 scope.go:117] "RemoveContainer" containerID="8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.941014 5035 scope.go:117] "RemoveContainer" containerID="ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264" Nov 24 21:14:45 crc kubenswrapper[5035]: E1124 21:14:45.941642 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264\": container with ID starting with ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264 not found: ID does not exist" containerID="ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.941672 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264"} err="failed to get container status \"ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264\": rpc error: code = NotFound desc = could not find container \"ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264\": container with ID starting with ae96c512de465e6d27fa28838289bc03c0b0c8c2fa52536eb8943d212eaf1264 not found: ID does not exist" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.941692 5035 scope.go:117] "RemoveContainer" containerID="3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae" Nov 24 21:14:45 crc kubenswrapper[5035]: E1124 21:14:45.941964 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae\": container with ID starting with 3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae not found: ID does not exist" containerID="3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.942004 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae"} err="failed to get container status \"3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae\": rpc error: code = NotFound desc = could not find container \"3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae\": container with ID starting with 3ca3314e66ab84dfe819dbaf4cace64fd0e7778fbf6f25615543678048ac3dae not found: ID does not exist" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.942025 5035 scope.go:117] "RemoveContainer" containerID="8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd" Nov 24 21:14:45 crc kubenswrapper[5035]: E1124 21:14:45.942242 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd\": container with ID starting with 8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd not found: ID does not exist" containerID="8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd" Nov 24 21:14:45 crc kubenswrapper[5035]: I1124 21:14:45.942272 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd"} err="failed to get container status \"8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd\": rpc error: code = NotFound desc = could not find container \"8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd\": container with ID starting with 8d34e9b57b1ae1092910e7254605074be85d1ba0b643fac6a6b4f3197ff858cd not found: ID does not exist" Nov 24 21:14:46 crc kubenswrapper[5035]: I1124 21:14:46.213879 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" path="/var/lib/kubelet/pods/80a3e8cf-9d55-48b6-9098-d4541e26d46b/volumes" Nov 24 21:14:46 crc kubenswrapper[5035]: I1124 21:14:46.793706 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzws2" event={"ID":"02e6808c-5f1e-4a32-bea6-bec389db20c2","Type":"ContainerStarted","Data":"c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5"} Nov 24 21:14:47 crc kubenswrapper[5035]: I1124 21:14:47.806001 5035 generic.go:334] "Generic (PLEG): container finished" podID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerID="c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5" exitCode=0 Nov 24 21:14:47 crc kubenswrapper[5035]: I1124 21:14:47.806077 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzws2" event={"ID":"02e6808c-5f1e-4a32-bea6-bec389db20c2","Type":"ContainerDied","Data":"c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5"} Nov 24 21:14:48 crc kubenswrapper[5035]: I1124 21:14:48.818125 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzws2" event={"ID":"02e6808c-5f1e-4a32-bea6-bec389db20c2","Type":"ContainerStarted","Data":"ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b"} Nov 24 21:14:48 crc kubenswrapper[5035]: I1124 21:14:48.840999 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jzws2" podStartSLOduration=2.433279008 podStartE2EDuration="4.840979029s" podCreationTimestamp="2025-11-24 21:14:44 +0000 UTC" firstStartedPulling="2025-11-24 21:14:45.806947114 +0000 UTC m=+7224.329453381" lastFinishedPulling="2025-11-24 21:14:48.214647145 +0000 UTC m=+7226.737153402" observedRunningTime="2025-11-24 21:14:48.837805954 +0000 UTC m=+7227.360312211" watchObservedRunningTime="2025-11-24 21:14:48.840979029 +0000 UTC m=+7227.363485286" Nov 24 21:14:50 crc kubenswrapper[5035]: I1124 21:14:50.043343 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:50 crc kubenswrapper[5035]: I1124 21:14:50.044541 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:50 crc kubenswrapper[5035]: I1124 21:14:50.101117 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:50 crc kubenswrapper[5035]: I1124 21:14:50.889328 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:52 crc kubenswrapper[5035]: I1124 21:14:52.503580 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzl4h"] Nov 24 21:14:52 crc kubenswrapper[5035]: I1124 21:14:52.855354 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qzl4h" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="registry-server" containerID="cri-o://ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6" gracePeriod=2 Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.365890 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.495365 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-utilities\") pod \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.495465 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-catalog-content\") pod \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.495557 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn4r5\" (UniqueName: \"kubernetes.io/projected/d27db1a1-f02b-4f19-a483-5aa0c79266dc-kube-api-access-sn4r5\") pod \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\" (UID: \"d27db1a1-f02b-4f19-a483-5aa0c79266dc\") " Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.498128 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-utilities" (OuterVolumeSpecName: "utilities") pod "d27db1a1-f02b-4f19-a483-5aa0c79266dc" (UID: "d27db1a1-f02b-4f19-a483-5aa0c79266dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.501028 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d27db1a1-f02b-4f19-a483-5aa0c79266dc-kube-api-access-sn4r5" (OuterVolumeSpecName: "kube-api-access-sn4r5") pod "d27db1a1-f02b-4f19-a483-5aa0c79266dc" (UID: "d27db1a1-f02b-4f19-a483-5aa0c79266dc"). InnerVolumeSpecName "kube-api-access-sn4r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.557137 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d27db1a1-f02b-4f19-a483-5aa0c79266dc" (UID: "d27db1a1-f02b-4f19-a483-5aa0c79266dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.597881 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.598108 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d27db1a1-f02b-4f19-a483-5aa0c79266dc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.598203 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn4r5\" (UniqueName: \"kubernetes.io/projected/d27db1a1-f02b-4f19-a483-5aa0c79266dc-kube-api-access-sn4r5\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.870634 5035 generic.go:334] "Generic (PLEG): container finished" podID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerID="ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6" exitCode=0 Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.870695 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerDied","Data":"ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6"} Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.870727 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzl4h" event={"ID":"d27db1a1-f02b-4f19-a483-5aa0c79266dc","Type":"ContainerDied","Data":"7455f04950a3e59a65010311097885948655ed6b1afdddc604e3809d2bc6465d"} Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.870749 5035 scope.go:117] "RemoveContainer" containerID="ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.870763 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzl4h" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.894034 5035 scope.go:117] "RemoveContainer" containerID="33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.918881 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzl4h"] Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.935092 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qzl4h"] Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.943347 5035 scope.go:117] "RemoveContainer" containerID="f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.973700 5035 scope.go:117] "RemoveContainer" containerID="ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6" Nov 24 21:14:53 crc kubenswrapper[5035]: E1124 21:14:53.974265 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6\": container with ID starting with ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6 not found: ID does not exist" containerID="ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.974444 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6"} err="failed to get container status \"ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6\": rpc error: code = NotFound desc = could not find container \"ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6\": container with ID starting with ad9ec5a7dcc37f1910be7bcacaeae89d192c2c8837d1fd1795a0aa99111936d6 not found: ID does not exist" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.974567 5035 scope.go:117] "RemoveContainer" containerID="33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5" Nov 24 21:14:53 crc kubenswrapper[5035]: E1124 21:14:53.975201 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5\": container with ID starting with 33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5 not found: ID does not exist" containerID="33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.975250 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5"} err="failed to get container status \"33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5\": rpc error: code = NotFound desc = could not find container \"33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5\": container with ID starting with 33041444e125f70cedf566b4980a864838e162b9c5ec98bec5df30f914654ef5 not found: ID does not exist" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.975277 5035 scope.go:117] "RemoveContainer" containerID="f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276" Nov 24 21:14:53 crc kubenswrapper[5035]: E1124 21:14:53.975612 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276\": container with ID starting with f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276 not found: ID does not exist" containerID="f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276" Nov 24 21:14:53 crc kubenswrapper[5035]: I1124 21:14:53.975656 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276"} err="failed to get container status \"f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276\": rpc error: code = NotFound desc = could not find container \"f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276\": container with ID starting with f6d5ada6d5e4969ebadfddb74c5ab4a54b266079cfcfa13dc7f33090aa7fb276 not found: ID does not exist" Nov 24 21:14:54 crc kubenswrapper[5035]: I1124 21:14:54.215517 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" path="/var/lib/kubelet/pods/d27db1a1-f02b-4f19-a483-5aa0c79266dc/volumes" Nov 24 21:14:54 crc kubenswrapper[5035]: I1124 21:14:54.856355 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:54 crc kubenswrapper[5035]: I1124 21:14:54.856410 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:54 crc kubenswrapper[5035]: I1124 21:14:54.913170 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:54 crc kubenswrapper[5035]: I1124 21:14:54.968165 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:56 crc kubenswrapper[5035]: I1124 21:14:56.200617 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:14:56 crc kubenswrapper[5035]: E1124 21:14:56.201203 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.302106 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzws2"] Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.302704 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jzws2" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="registry-server" containerID="cri-o://ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b" gracePeriod=2 Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.893468 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.915867 5035 generic.go:334] "Generic (PLEG): container finished" podID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerID="ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b" exitCode=0 Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.915920 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzws2" event={"ID":"02e6808c-5f1e-4a32-bea6-bec389db20c2","Type":"ContainerDied","Data":"ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b"} Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.915953 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzws2" event={"ID":"02e6808c-5f1e-4a32-bea6-bec389db20c2","Type":"ContainerDied","Data":"39e4333863ea6806e72e11bb024349295dddd7825e4e5e9cc3e746d5dcdc818c"} Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.915974 5035 scope.go:117] "RemoveContainer" containerID="ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b" Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.916141 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzws2" Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.972508 5035 scope.go:117] "RemoveContainer" containerID="c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5" Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.987838 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-utilities\") pod \"02e6808c-5f1e-4a32-bea6-bec389db20c2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.987996 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldnzs\" (UniqueName: \"kubernetes.io/projected/02e6808c-5f1e-4a32-bea6-bec389db20c2-kube-api-access-ldnzs\") pod \"02e6808c-5f1e-4a32-bea6-bec389db20c2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.988091 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-catalog-content\") pod \"02e6808c-5f1e-4a32-bea6-bec389db20c2\" (UID: \"02e6808c-5f1e-4a32-bea6-bec389db20c2\") " Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.990552 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-utilities" (OuterVolumeSpecName: "utilities") pod "02e6808c-5f1e-4a32-bea6-bec389db20c2" (UID: "02e6808c-5f1e-4a32-bea6-bec389db20c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.996067 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e6808c-5f1e-4a32-bea6-bec389db20c2-kube-api-access-ldnzs" (OuterVolumeSpecName: "kube-api-access-ldnzs") pod "02e6808c-5f1e-4a32-bea6-bec389db20c2" (UID: "02e6808c-5f1e-4a32-bea6-bec389db20c2"). InnerVolumeSpecName "kube-api-access-ldnzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:14:57 crc kubenswrapper[5035]: I1124 21:14:57.996415 5035 scope.go:117] "RemoveContainer" containerID="97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.012830 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02e6808c-5f1e-4a32-bea6-bec389db20c2" (UID: "02e6808c-5f1e-4a32-bea6-bec389db20c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.085010 5035 scope.go:117] "RemoveContainer" containerID="ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b" Nov 24 21:14:58 crc kubenswrapper[5035]: E1124 21:14:58.085483 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b\": container with ID starting with ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b not found: ID does not exist" containerID="ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.085530 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b"} err="failed to get container status \"ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b\": rpc error: code = NotFound desc = could not find container \"ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b\": container with ID starting with ae4468fecdd64e5258e80633f3e846906326d2fd0bfe298dd913f05e1de5041b not found: ID does not exist" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.085557 5035 scope.go:117] "RemoveContainer" containerID="c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5" Nov 24 21:14:58 crc kubenswrapper[5035]: E1124 21:14:58.085927 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5\": container with ID starting with c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5 not found: ID does not exist" containerID="c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.085966 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5"} err="failed to get container status \"c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5\": rpc error: code = NotFound desc = could not find container \"c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5\": container with ID starting with c4ac582278f53f43a9ed8e14232a99c77c89fdbd1d856e5192050d793106fdf5 not found: ID does not exist" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.086009 5035 scope.go:117] "RemoveContainer" containerID="97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121" Nov 24 21:14:58 crc kubenswrapper[5035]: E1124 21:14:58.086257 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121\": container with ID starting with 97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121 not found: ID does not exist" containerID="97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.086280 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121"} err="failed to get container status \"97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121\": rpc error: code = NotFound desc = could not find container \"97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121\": container with ID starting with 97eeda6a6a719dd8e6f609d4948d28a49c410120a6323be7ed1c3006ef4b6121 not found: ID does not exist" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.091115 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.091141 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e6808c-5f1e-4a32-bea6-bec389db20c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.091153 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldnzs\" (UniqueName: \"kubernetes.io/projected/02e6808c-5f1e-4a32-bea6-bec389db20c2-kube-api-access-ldnzs\") on node \"crc\" DevicePath \"\"" Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.263741 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzws2"] Nov 24 21:14:58 crc kubenswrapper[5035]: I1124 21:14:58.275032 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzws2"] Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.158723 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh"] Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159510 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="extract-content" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159539 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="extract-content" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159553 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159561 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159569 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="extract-content" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159575 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="extract-content" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159583 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="extract-utilities" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159589 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="extract-utilities" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159597 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159603 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159612 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="extract-content" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159619 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="extract-content" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159630 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="extract-utilities" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159636 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="extract-utilities" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159677 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159684 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: E1124 21:15:00.159693 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="extract-utilities" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159699 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="extract-utilities" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159877 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159888 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d27db1a1-f02b-4f19-a483-5aa0c79266dc" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.159901 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="80a3e8cf-9d55-48b6-9098-d4541e26d46b" containerName="registry-server" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.160689 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.164115 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.164125 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.175892 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh"] Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.214997 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02e6808c-5f1e-4a32-bea6-bec389db20c2" path="/var/lib/kubelet/pods/02e6808c-5f1e-4a32-bea6-bec389db20c2/volumes" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.329050 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f972291-8bf4-49c7-83ea-7cbb858ec89c-secret-volume\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.329195 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwpst\" (UniqueName: \"kubernetes.io/projected/7f972291-8bf4-49c7-83ea-7cbb858ec89c-kube-api-access-lwpst\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.329246 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f972291-8bf4-49c7-83ea-7cbb858ec89c-config-volume\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.431447 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f972291-8bf4-49c7-83ea-7cbb858ec89c-secret-volume\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.431886 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwpst\" (UniqueName: \"kubernetes.io/projected/7f972291-8bf4-49c7-83ea-7cbb858ec89c-kube-api-access-lwpst\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.431950 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f972291-8bf4-49c7-83ea-7cbb858ec89c-config-volume\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.433183 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f972291-8bf4-49c7-83ea-7cbb858ec89c-config-volume\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.446942 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f972291-8bf4-49c7-83ea-7cbb858ec89c-secret-volume\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.458009 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwpst\" (UniqueName: \"kubernetes.io/projected/7f972291-8bf4-49c7-83ea-7cbb858ec89c-kube-api-access-lwpst\") pod \"collect-profiles-29400315-7cbfh\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:00 crc kubenswrapper[5035]: I1124 21:15:00.507855 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:01 crc kubenswrapper[5035]: I1124 21:15:01.821643 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh"] Nov 24 21:15:01 crc kubenswrapper[5035]: I1124 21:15:01.959195 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" event={"ID":"7f972291-8bf4-49c7-83ea-7cbb858ec89c","Type":"ContainerStarted","Data":"318dce8c1c90915c476bc53169b385d499fc368e851f2ec928dda179d06d42be"} Nov 24 21:15:02 crc kubenswrapper[5035]: I1124 21:15:02.976764 5035 generic.go:334] "Generic (PLEG): container finished" podID="7f972291-8bf4-49c7-83ea-7cbb858ec89c" containerID="955946576ea65bf08c502ba2f35e98c0b7dba3a678e361d47263928d28d0cf99" exitCode=0 Nov 24 21:15:02 crc kubenswrapper[5035]: I1124 21:15:02.976816 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" event={"ID":"7f972291-8bf4-49c7-83ea-7cbb858ec89c","Type":"ContainerDied","Data":"955946576ea65bf08c502ba2f35e98c0b7dba3a678e361d47263928d28d0cf99"} Nov 24 21:15:02 crc kubenswrapper[5035]: I1124 21:15:02.979151 5035 generic.go:334] "Generic (PLEG): container finished" podID="e786b372-cf0a-4004-b6a3-0a13e6fc66e8" containerID="7224554783c82c0a2b63712d7a321740192e0bff29db812e4c752143b003871f" exitCode=0 Nov 24 21:15:02 crc kubenswrapper[5035]: I1124 21:15:02.979359 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" event={"ID":"e786b372-cf0a-4004-b6a3-0a13e6fc66e8","Type":"ContainerDied","Data":"7224554783c82c0a2b63712d7a321740192e0bff29db812e4c752143b003871f"} Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.087900 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.126529 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-b95zr"] Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.135696 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-b95zr"] Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.211780 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxgk7\" (UniqueName: \"kubernetes.io/projected/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-kube-api-access-kxgk7\") pod \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.211850 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-host\") pod \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\" (UID: \"e786b372-cf0a-4004-b6a3-0a13e6fc66e8\") " Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.211901 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-host" (OuterVolumeSpecName: "host") pod "e786b372-cf0a-4004-b6a3-0a13e6fc66e8" (UID: "e786b372-cf0a-4004-b6a3-0a13e6fc66e8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.212884 5035 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-host\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.220086 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-kube-api-access-kxgk7" (OuterVolumeSpecName: "kube-api-access-kxgk7") pod "e786b372-cf0a-4004-b6a3-0a13e6fc66e8" (UID: "e786b372-cf0a-4004-b6a3-0a13e6fc66e8"). InnerVolumeSpecName "kube-api-access-kxgk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.315392 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxgk7\" (UniqueName: \"kubernetes.io/projected/e786b372-cf0a-4004-b6a3-0a13e6fc66e8-kube-api-access-kxgk7\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.334589 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.518189 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f972291-8bf4-49c7-83ea-7cbb858ec89c-config-volume\") pod \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.518243 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f972291-8bf4-49c7-83ea-7cbb858ec89c-secret-volume\") pod \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.518427 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwpst\" (UniqueName: \"kubernetes.io/projected/7f972291-8bf4-49c7-83ea-7cbb858ec89c-kube-api-access-lwpst\") pod \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\" (UID: \"7f972291-8bf4-49c7-83ea-7cbb858ec89c\") " Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.519137 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f972291-8bf4-49c7-83ea-7cbb858ec89c-config-volume" (OuterVolumeSpecName: "config-volume") pod "7f972291-8bf4-49c7-83ea-7cbb858ec89c" (UID: "7f972291-8bf4-49c7-83ea-7cbb858ec89c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.523319 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f972291-8bf4-49c7-83ea-7cbb858ec89c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7f972291-8bf4-49c7-83ea-7cbb858ec89c" (UID: "7f972291-8bf4-49c7-83ea-7cbb858ec89c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.527904 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f972291-8bf4-49c7-83ea-7cbb858ec89c-kube-api-access-lwpst" (OuterVolumeSpecName: "kube-api-access-lwpst") pod "7f972291-8bf4-49c7-83ea-7cbb858ec89c" (UID: "7f972291-8bf4-49c7-83ea-7cbb858ec89c"). InnerVolumeSpecName "kube-api-access-lwpst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.621314 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f972291-8bf4-49c7-83ea-7cbb858ec89c-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.621534 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f972291-8bf4-49c7-83ea-7cbb858ec89c-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.621548 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwpst\" (UniqueName: \"kubernetes.io/projected/7f972291-8bf4-49c7-83ea-7cbb858ec89c-kube-api-access-lwpst\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.998900 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" event={"ID":"7f972291-8bf4-49c7-83ea-7cbb858ec89c","Type":"ContainerDied","Data":"318dce8c1c90915c476bc53169b385d499fc368e851f2ec928dda179d06d42be"} Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.998938 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="318dce8c1c90915c476bc53169b385d499fc368e851f2ec928dda179d06d42be" Nov 24 21:15:04 crc kubenswrapper[5035]: I1124 21:15:04.998939 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400315-7cbfh" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.009228 5035 scope.go:117] "RemoveContainer" containerID="7224554783c82c0a2b63712d7a321740192e0bff29db812e4c752143b003871f" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.009455 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-b95zr" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.332070 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-j2q2g"] Nov 24 21:15:05 crc kubenswrapper[5035]: E1124 21:15:05.332527 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e786b372-cf0a-4004-b6a3-0a13e6fc66e8" containerName="container-00" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.332543 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e786b372-cf0a-4004-b6a3-0a13e6fc66e8" containerName="container-00" Nov 24 21:15:05 crc kubenswrapper[5035]: E1124 21:15:05.332579 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f972291-8bf4-49c7-83ea-7cbb858ec89c" containerName="collect-profiles" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.332586 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f972291-8bf4-49c7-83ea-7cbb858ec89c" containerName="collect-profiles" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.332818 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e786b372-cf0a-4004-b6a3-0a13e6fc66e8" containerName="container-00" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.332837 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f972291-8bf4-49c7-83ea-7cbb858ec89c" containerName="collect-profiles" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.333626 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.404155 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t"] Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.412084 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400270-qpb4t"] Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.437014 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/389a4fa7-9685-442a-b3c2-f1b264a4c34e-host\") pod \"crc-debug-j2q2g\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.437239 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch7hd\" (UniqueName: \"kubernetes.io/projected/389a4fa7-9685-442a-b3c2-f1b264a4c34e-kube-api-access-ch7hd\") pod \"crc-debug-j2q2g\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.539501 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/389a4fa7-9685-442a-b3c2-f1b264a4c34e-host\") pod \"crc-debug-j2q2g\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.539555 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch7hd\" (UniqueName: \"kubernetes.io/projected/389a4fa7-9685-442a-b3c2-f1b264a4c34e-kube-api-access-ch7hd\") pod \"crc-debug-j2q2g\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.540034 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/389a4fa7-9685-442a-b3c2-f1b264a4c34e-host\") pod \"crc-debug-j2q2g\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.558227 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch7hd\" (UniqueName: \"kubernetes.io/projected/389a4fa7-9685-442a-b3c2-f1b264a4c34e-kube-api-access-ch7hd\") pod \"crc-debug-j2q2g\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: I1124 21:15:05.652120 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:05 crc kubenswrapper[5035]: W1124 21:15:05.678384 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod389a4fa7_9685_442a_b3c2_f1b264a4c34e.slice/crio-4b36f75bc63110f88904d645e8b7b2c9eea349033b519f512fc4ea64b339715f WatchSource:0}: Error finding container 4b36f75bc63110f88904d645e8b7b2c9eea349033b519f512fc4ea64b339715f: Status 404 returned error can't find the container with id 4b36f75bc63110f88904d645e8b7b2c9eea349033b519f512fc4ea64b339715f Nov 24 21:15:06 crc kubenswrapper[5035]: I1124 21:15:06.021979 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" event={"ID":"389a4fa7-9685-442a-b3c2-f1b264a4c34e","Type":"ContainerStarted","Data":"0af64bac3ebd609f82b9f250f5140b9d7409d8857f0a98c6b2be6457bd438f5c"} Nov 24 21:15:06 crc kubenswrapper[5035]: I1124 21:15:06.022219 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" event={"ID":"389a4fa7-9685-442a-b3c2-f1b264a4c34e","Type":"ContainerStarted","Data":"4b36f75bc63110f88904d645e8b7b2c9eea349033b519f512fc4ea64b339715f"} Nov 24 21:15:06 crc kubenswrapper[5035]: I1124 21:15:06.040972 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" podStartSLOduration=1.040956937 podStartE2EDuration="1.040956937s" podCreationTimestamp="2025-11-24 21:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 21:15:06.039498088 +0000 UTC m=+7244.562004345" watchObservedRunningTime="2025-11-24 21:15:06.040956937 +0000 UTC m=+7244.563463194" Nov 24 21:15:06 crc kubenswrapper[5035]: I1124 21:15:06.211986 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e786b372-cf0a-4004-b6a3-0a13e6fc66e8" path="/var/lib/kubelet/pods/e786b372-cf0a-4004-b6a3-0a13e6fc66e8/volumes" Nov 24 21:15:06 crc kubenswrapper[5035]: I1124 21:15:06.212620 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9d1d182-8018-4fe4-804f-aa198fdf4812" path="/var/lib/kubelet/pods/e9d1d182-8018-4fe4-804f-aa198fdf4812/volumes" Nov 24 21:15:07 crc kubenswrapper[5035]: I1124 21:15:07.032870 5035 generic.go:334] "Generic (PLEG): container finished" podID="389a4fa7-9685-442a-b3c2-f1b264a4c34e" containerID="0af64bac3ebd609f82b9f250f5140b9d7409d8857f0a98c6b2be6457bd438f5c" exitCode=0 Nov 24 21:15:07 crc kubenswrapper[5035]: I1124 21:15:07.032981 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" event={"ID":"389a4fa7-9685-442a-b3c2-f1b264a4c34e","Type":"ContainerDied","Data":"0af64bac3ebd609f82b9f250f5140b9d7409d8857f0a98c6b2be6457bd438f5c"} Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.151242 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.284511 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch7hd\" (UniqueName: \"kubernetes.io/projected/389a4fa7-9685-442a-b3c2-f1b264a4c34e-kube-api-access-ch7hd\") pod \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.285798 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/389a4fa7-9685-442a-b3c2-f1b264a4c34e-host\") pod \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\" (UID: \"389a4fa7-9685-442a-b3c2-f1b264a4c34e\") " Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.285953 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/389a4fa7-9685-442a-b3c2-f1b264a4c34e-host" (OuterVolumeSpecName: "host") pod "389a4fa7-9685-442a-b3c2-f1b264a4c34e" (UID: "389a4fa7-9685-442a-b3c2-f1b264a4c34e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.287798 5035 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/389a4fa7-9685-442a-b3c2-f1b264a4c34e-host\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.295270 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/389a4fa7-9685-442a-b3c2-f1b264a4c34e-kube-api-access-ch7hd" (OuterVolumeSpecName: "kube-api-access-ch7hd") pod "389a4fa7-9685-442a-b3c2-f1b264a4c34e" (UID: "389a4fa7-9685-442a-b3c2-f1b264a4c34e"). InnerVolumeSpecName "kube-api-access-ch7hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.389028 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch7hd\" (UniqueName: \"kubernetes.io/projected/389a4fa7-9685-442a-b3c2-f1b264a4c34e-kube-api-access-ch7hd\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.856340 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-j2q2g"] Nov 24 21:15:08 crc kubenswrapper[5035]: I1124 21:15:08.864072 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-j2q2g"] Nov 24 21:15:09 crc kubenswrapper[5035]: I1124 21:15:09.052481 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b36f75bc63110f88904d645e8b7b2c9eea349033b519f512fc4ea64b339715f" Nov 24 21:15:09 crc kubenswrapper[5035]: I1124 21:15:09.052548 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-j2q2g" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.012705 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-v6kf4"] Nov 24 21:15:10 crc kubenswrapper[5035]: E1124 21:15:10.013388 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="389a4fa7-9685-442a-b3c2-f1b264a4c34e" containerName="container-00" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.013404 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="389a4fa7-9685-442a-b3c2-f1b264a4c34e" containerName="container-00" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.013591 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="389a4fa7-9685-442a-b3c2-f1b264a4c34e" containerName="container-00" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.014322 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.121582 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e750574-8891-4edc-b063-b3952b3f42d7-host\") pod \"crc-debug-v6kf4\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.121639 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6knws\" (UniqueName: \"kubernetes.io/projected/0e750574-8891-4edc-b063-b3952b3f42d7-kube-api-access-6knws\") pod \"crc-debug-v6kf4\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.200279 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:15:10 crc kubenswrapper[5035]: E1124 21:15:10.200614 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.210865 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="389a4fa7-9685-442a-b3c2-f1b264a4c34e" path="/var/lib/kubelet/pods/389a4fa7-9685-442a-b3c2-f1b264a4c34e/volumes" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.224072 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e750574-8891-4edc-b063-b3952b3f42d7-host\") pod \"crc-debug-v6kf4\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.224137 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6knws\" (UniqueName: \"kubernetes.io/projected/0e750574-8891-4edc-b063-b3952b3f42d7-kube-api-access-6knws\") pod \"crc-debug-v6kf4\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.224151 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e750574-8891-4edc-b063-b3952b3f42d7-host\") pod \"crc-debug-v6kf4\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.249209 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6knws\" (UniqueName: \"kubernetes.io/projected/0e750574-8891-4edc-b063-b3952b3f42d7-kube-api-access-6knws\") pod \"crc-debug-v6kf4\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:10 crc kubenswrapper[5035]: I1124 21:15:10.332201 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:11 crc kubenswrapper[5035]: I1124 21:15:11.075357 5035 generic.go:334] "Generic (PLEG): container finished" podID="0e750574-8891-4edc-b063-b3952b3f42d7" containerID="b105eec99891f50f60941b14f545682b1fa4dd6ff10990de61b49d0f47cf1de4" exitCode=0 Nov 24 21:15:11 crc kubenswrapper[5035]: I1124 21:15:11.075475 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" event={"ID":"0e750574-8891-4edc-b063-b3952b3f42d7","Type":"ContainerDied","Data":"b105eec99891f50f60941b14f545682b1fa4dd6ff10990de61b49d0f47cf1de4"} Nov 24 21:15:11 crc kubenswrapper[5035]: I1124 21:15:11.075620 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" event={"ID":"0e750574-8891-4edc-b063-b3952b3f42d7","Type":"ContainerStarted","Data":"203e21f7e988899a2b9973191fcde888860a3e5879cce3a11ad77e6c0e3def53"} Nov 24 21:15:11 crc kubenswrapper[5035]: I1124 21:15:11.116364 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-v6kf4"] Nov 24 21:15:11 crc kubenswrapper[5035]: I1124 21:15:11.126697 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pw2tc/crc-debug-v6kf4"] Nov 24 21:15:12 crc kubenswrapper[5035]: I1124 21:15:12.174611 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:12 crc kubenswrapper[5035]: I1124 21:15:12.270612 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6knws\" (UniqueName: \"kubernetes.io/projected/0e750574-8891-4edc-b063-b3952b3f42d7-kube-api-access-6knws\") pod \"0e750574-8891-4edc-b063-b3952b3f42d7\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " Nov 24 21:15:12 crc kubenswrapper[5035]: I1124 21:15:12.271167 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e750574-8891-4edc-b063-b3952b3f42d7-host\") pod \"0e750574-8891-4edc-b063-b3952b3f42d7\" (UID: \"0e750574-8891-4edc-b063-b3952b3f42d7\") " Nov 24 21:15:12 crc kubenswrapper[5035]: I1124 21:15:12.271226 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e750574-8891-4edc-b063-b3952b3f42d7-host" (OuterVolumeSpecName: "host") pod "0e750574-8891-4edc-b063-b3952b3f42d7" (UID: "0e750574-8891-4edc-b063-b3952b3f42d7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 21:15:12 crc kubenswrapper[5035]: I1124 21:15:12.272001 5035 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e750574-8891-4edc-b063-b3952b3f42d7-host\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:12 crc kubenswrapper[5035]: I1124 21:15:12.278813 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e750574-8891-4edc-b063-b3952b3f42d7-kube-api-access-6knws" (OuterVolumeSpecName: "kube-api-access-6knws") pod "0e750574-8891-4edc-b063-b3952b3f42d7" (UID: "0e750574-8891-4edc-b063-b3952b3f42d7"). InnerVolumeSpecName "kube-api-access-6knws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:15:12 crc kubenswrapper[5035]: I1124 21:15:12.374161 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6knws\" (UniqueName: \"kubernetes.io/projected/0e750574-8891-4edc-b063-b3952b3f42d7-kube-api-access-6knws\") on node \"crc\" DevicePath \"\"" Nov 24 21:15:13 crc kubenswrapper[5035]: I1124 21:15:13.096570 5035 scope.go:117] "RemoveContainer" containerID="b105eec99891f50f60941b14f545682b1fa4dd6ff10990de61b49d0f47cf1de4" Nov 24 21:15:13 crc kubenswrapper[5035]: I1124 21:15:13.096667 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/crc-debug-v6kf4" Nov 24 21:15:14 crc kubenswrapper[5035]: I1124 21:15:14.212732 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e750574-8891-4edc-b063-b3952b3f42d7" path="/var/lib/kubelet/pods/0e750574-8891-4edc-b063-b3952b3f42d7/volumes" Nov 24 21:15:24 crc kubenswrapper[5035]: I1124 21:15:24.201447 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:15:24 crc kubenswrapper[5035]: E1124 21:15:24.202512 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:15:33 crc kubenswrapper[5035]: I1124 21:15:33.587234 5035 scope.go:117] "RemoveContainer" containerID="6f579b2454f0b077d89c1ab3db50d1ba0358791295b7274b4c22594ab613587c" Nov 24 21:15:33 crc kubenswrapper[5035]: I1124 21:15:33.998191 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_0eab5718-a1b7-4621-96bd-b35f01b4ed48/ansibletest-ansibletest/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.138507 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-555d68d8cb-knjvm_39154cf4-dfd5-4fb2-bd13-e9d15012da12/barbican-api/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.160487 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-555d68d8cb-knjvm_39154cf4-dfd5-4fb2-bd13-e9d15012da12/barbican-api-log/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.296328 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-bf699cb66-l5qkm_85943038-29af-4153-be73-f62fb83bf015/barbican-keystone-listener/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.493083 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9df689fb7-8fbwv_561f02ab-04d0-4100-8e25-66fd0df630f6/barbican-worker/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.569678 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-9df689fb7-8fbwv_561f02ab-04d0-4100-8e25-66fd0df630f6/barbican-worker-log/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.734606 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-jvtf9_8f7443fa-cea8-4010-8cac-0186257581cb/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.891264 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f9d5829b-2574-4340-aa87-c7d78011d378/ceilometer-central-agent/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.906536 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-bf699cb66-l5qkm_85943038-29af-4153-be73-f62fb83bf015/barbican-keystone-listener-log/0.log" Nov 24 21:15:34 crc kubenswrapper[5035]: I1124 21:15:34.987595 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f9d5829b-2574-4340-aa87-c7d78011d378/ceilometer-notification-agent/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.085031 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f9d5829b-2574-4340-aa87-c7d78011d378/proxy-httpd/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.096547 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f9d5829b-2574-4340-aa87-c7d78011d378/sg-core/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.189332 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-ntnxb_749b308d-659f-4128-b53c-cefd8ab3e1d6/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.285056 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lrr9x_6ef15736-c842-4a20-b05f-9105eef653a7/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.481494 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a37be962-03c8-4e0a-b329-12d6765d5c77/cinder-api/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.520967 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a37be962-03c8-4e0a-b329-12d6765d5c77/cinder-api-log/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.731174 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_fc314ad6-dfe5-4703-9f08-f08f904b605c/probe/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.753161 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_fc314ad6-dfe5-4703-9f08-f08f904b605c/cinder-backup/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.820490 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a1d160b0-7b77-4f8c-a918-170911d5ca2a/cinder-scheduler/0.log" Nov 24 21:15:35 crc kubenswrapper[5035]: I1124 21:15:35.972894 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a1d160b0-7b77-4f8c-a918-170911d5ca2a/probe/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.050995 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_4d0bc83b-3e60-45a1-a251-06a00909de0a/cinder-volume/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.074493 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_4d0bc83b-3e60-45a1-a251-06a00909de0a/probe/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.419343 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zfqq7_8eb8573f-9078-4348-a9e3-d4156a32a917/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.511401 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-j4cdn_d0ebd40d-a3e7-4358-93be-983c52b4f89d/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.658016 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78f48d6b7c-zbdgc_8d6b76db-f778-4b7e-9dab-e0fa5921458e/init/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.879157 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78f48d6b7c-zbdgc_8d6b76db-f778-4b7e-9dab-e0fa5921458e/dnsmasq-dns/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.914838 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_815bccaf-12a3-47b8-8e80-3c1c30724697/glance-httpd/0.log" Nov 24 21:15:36 crc kubenswrapper[5035]: I1124 21:15:36.921436 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78f48d6b7c-zbdgc_8d6b76db-f778-4b7e-9dab-e0fa5921458e/init/0.log" Nov 24 21:15:37 crc kubenswrapper[5035]: I1124 21:15:37.123230 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_815bccaf-12a3-47b8-8e80-3c1c30724697/glance-log/0.log" Nov 24 21:15:37 crc kubenswrapper[5035]: I1124 21:15:37.194895 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e7851793-a562-491b-8c0e-47b9155a75bc/glance-httpd/0.log" Nov 24 21:15:37 crc kubenswrapper[5035]: I1124 21:15:37.200557 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e7851793-a562-491b-8c0e-47b9155a75bc/glance-log/0.log" Nov 24 21:15:37 crc kubenswrapper[5035]: I1124 21:15:37.477432 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-677648b7f8-j8p2d_6c39b1fa-70b7-49f4-b318-0d08d5007dba/horizon/0.log" Nov 24 21:15:37 crc kubenswrapper[5035]: I1124 21:15:37.627914 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_36ff01ed-eaf8-488f-9243-44710cb7017f/horizontest-tests-horizontest/0.log" Nov 24 21:15:37 crc kubenswrapper[5035]: I1124 21:15:37.766457 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-m8xt4_bae54c1b-690d-4e84-8c7e-6f4fdab2ae26/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:37 crc kubenswrapper[5035]: I1124 21:15:37.950142 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-klf6r_e6f5bc3a-66aa-4f6a-b098-7b6a2512f120/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.196610 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400241-gqccc_549314e7-60d0-4aa6-a48b-a36acdaefa29/keystone-cron/0.log" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.202932 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:15:38 crc kubenswrapper[5035]: E1124 21:15:38.203166 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.400415 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400301-dxpzw_46fd0927-55d9-4806-b7ad-2d3633aad62a/keystone-cron/0.log" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.477137 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_bc86b831-9dce-4436-a420-a2f14364bea4/kube-state-metrics/0.log" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.588427 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-677648b7f8-j8p2d_6c39b1fa-70b7-49f4-b318-0d08d5007dba/horizon-log/0.log" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.782388 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-fj957_db87bf3b-6dd8-448d-80fe-8b8db1ecca5d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.904151 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_c994550c-82c0-4e4a-9587-0493a16703e7/manila-api-log/0.log" Nov 24 21:15:38 crc kubenswrapper[5035]: I1124 21:15:38.956180 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_c994550c-82c0-4e4a-9587-0493a16703e7/manila-api/0.log" Nov 24 21:15:39 crc kubenswrapper[5035]: E1124 21:15:39.200355 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:15:39 crc kubenswrapper[5035]: I1124 21:15:39.279374 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3beb7c92-aa3c-425d-9964-1dfa44680669/manila-scheduler/0.log" Nov 24 21:15:39 crc kubenswrapper[5035]: I1124 21:15:39.317759 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c9499d9cb-dtnz8_cc8d3b90-b7f5-468b-8a8e-7c75b4afd1ae/keystone-api/0.log" Nov 24 21:15:39 crc kubenswrapper[5035]: I1124 21:15:39.338038 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3beb7c92-aa3c-425d-9964-1dfa44680669/probe/0.log" Nov 24 21:15:39 crc kubenswrapper[5035]: I1124 21:15:39.515900 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_53e5b6e3-1cf1-4537-8263-553481eb0242/probe/0.log" Nov 24 21:15:39 crc kubenswrapper[5035]: I1124 21:15:39.527094 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_53e5b6e3-1cf1-4537-8263-553481eb0242/manila-share/0.log" Nov 24 21:15:39 crc kubenswrapper[5035]: I1124 21:15:39.911055 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-6kp48_ecfdf321-e4fd-4703-a5cd-907d6508d7e6/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:40 crc kubenswrapper[5035]: I1124 21:15:40.200114 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b7c585949-9h76s_6886d0f1-8b8b-4580-8145-f5bb671cdf1f/neutron-httpd/0.log" Nov 24 21:15:40 crc kubenswrapper[5035]: I1124 21:15:40.852929 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b7c585949-9h76s_6886d0f1-8b8b-4580-8145-f5bb671cdf1f/neutron-api/0.log" Nov 24 21:15:41 crc kubenswrapper[5035]: I1124 21:15:41.548898 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5a4dd86f-4c97-4f2d-b67b-ce7746d105ca/nova-cell1-conductor-conductor/0.log" Nov 24 21:15:41 crc kubenswrapper[5035]: I1124 21:15:41.719007 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_817624af-1239-4775-b5d2-87d0c0b7ef95/nova-cell0-conductor-conductor/0.log" Nov 24 21:15:42 crc kubenswrapper[5035]: I1124 21:15:42.102965 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6a38d999-4d5c-47ad-91af-605b01a6daab/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 21:15:42 crc kubenswrapper[5035]: I1124 21:15:42.334327 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8fqkx_4bbedfca-6249-484a-a6cf-da8b54a51b1e/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:42 crc kubenswrapper[5035]: I1124 21:15:42.643153 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5e338747-46ee-4e25-a2cc-6696a162487a/nova-metadata-log/0.log" Nov 24 21:15:43 crc kubenswrapper[5035]: I1124 21:15:43.080698 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2f870b4b-afbf-48bc-95d0-ba675cffaa49/nova-api-log/0.log" Nov 24 21:15:43 crc kubenswrapper[5035]: I1124 21:15:43.491168 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_62cb1b72-4c4f-4c1e-bf23-0dc8a318560d/mysql-bootstrap/0.log" Nov 24 21:15:43 crc kubenswrapper[5035]: I1124 21:15:43.641637 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_5bd81445-025b-4359-8783-2da4b4181807/nova-scheduler-scheduler/0.log" Nov 24 21:15:43 crc kubenswrapper[5035]: I1124 21:15:43.680822 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_62cb1b72-4c4f-4c1e-bf23-0dc8a318560d/mysql-bootstrap/0.log" Nov 24 21:15:43 crc kubenswrapper[5035]: I1124 21:15:43.807077 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_62cb1b72-4c4f-4c1e-bf23-0dc8a318560d/galera/0.log" Nov 24 21:15:43 crc kubenswrapper[5035]: I1124 21:15:43.899142 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2f870b4b-afbf-48bc-95d0-ba675cffaa49/nova-api-api/0.log" Nov 24 21:15:44 crc kubenswrapper[5035]: I1124 21:15:44.032838 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_fcbe1807-d08b-4d8a-9c92-954b8da6820d/mysql-bootstrap/0.log" Nov 24 21:15:44 crc kubenswrapper[5035]: I1124 21:15:44.259577 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_fcbe1807-d08b-4d8a-9c92-954b8da6820d/mysql-bootstrap/0.log" Nov 24 21:15:44 crc kubenswrapper[5035]: I1124 21:15:44.264159 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_fcbe1807-d08b-4d8a-9c92-954b8da6820d/galera/0.log" Nov 24 21:15:44 crc kubenswrapper[5035]: I1124 21:15:44.510859 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1128a14c-1553-433f-b64b-676f3cec11f9/openstackclient/0.log" Nov 24 21:15:44 crc kubenswrapper[5035]: I1124 21:15:44.595454 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-82r49_aa0e4fdb-6066-4163-9864-9e52443afdb0/ovn-controller/0.log" Nov 24 21:15:44 crc kubenswrapper[5035]: I1124 21:15:44.739718 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rxj4x_2a4176a5-5ca8-4a71-8cab-3330b6af69f5/openstack-network-exporter/0.log" Nov 24 21:15:44 crc kubenswrapper[5035]: I1124 21:15:44.905253 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6l2z4_89539b2e-806f-4062-85fe-efac9c4469d2/ovsdb-server-init/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.202857 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6l2z4_89539b2e-806f-4062-85fe-efac9c4469d2/ovs-vswitchd/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.214856 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6l2z4_89539b2e-806f-4062-85fe-efac9c4469d2/ovsdb-server/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.232737 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6l2z4_89539b2e-806f-4062-85fe-efac9c4469d2/ovsdb-server-init/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.451579 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-b4rlz_f2273d9c-4343-4bb5-9bca-0a87c601cf15/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.663500 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9637335d-6e3c-4f1a-bfcd-a081f20e8226/openstack-network-exporter/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.669187 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9637335d-6e3c-4f1a-bfcd-a081f20e8226/ovn-northd/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.803969 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e665c207-0fb4-4876-89d1-cb18eadb80af/memcached/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.863314 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_74f6b29d-0844-4946-abaf-d331f8f07ba0/openstack-network-exporter/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.882718 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_74f6b29d-0844-4946-abaf-d331f8f07ba0/ovsdbserver-nb/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.888169 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5e338747-46ee-4e25-a2cc-6696a162487a/nova-metadata-metadata/0.log" Nov 24 21:15:45 crc kubenswrapper[5035]: I1124 21:15:45.985181 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf/openstack-network-exporter/0.log" Nov 24 21:15:46 crc kubenswrapper[5035]: I1124 21:15:46.057848 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1ddbd9ab-f1f3-4243-be2d-6653c8dca6cf/ovsdbserver-sb/0.log" Nov 24 21:15:46 crc kubenswrapper[5035]: I1124 21:15:46.246565 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6d6f3cf9-b4c1-48e1-880b-8812b7f4421c/setup-container/0.log" Nov 24 21:15:46 crc kubenswrapper[5035]: I1124 21:15:46.478572 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6869cb8db6-rk7hs_f9f396db-4624-4c15-91d5-47c22dbeb77c/placement-api/0.log" Nov 24 21:15:46 crc kubenswrapper[5035]: I1124 21:15:46.490121 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6d6f3cf9-b4c1-48e1-880b-8812b7f4421c/setup-container/0.log" Nov 24 21:15:46 crc kubenswrapper[5035]: I1124 21:15:46.555576 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6d6f3cf9-b4c1-48e1-880b-8812b7f4421c/rabbitmq/0.log" Nov 24 21:15:46 crc kubenswrapper[5035]: I1124 21:15:46.555711 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6869cb8db6-rk7hs_f9f396db-4624-4c15-91d5-47c22dbeb77c/placement-log/0.log" Nov 24 21:15:46 crc kubenswrapper[5035]: I1124 21:15:46.902418 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_00618566-3b37-4b50-92d5-5f7dabf6a55b/setup-container/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.027720 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_00618566-3b37-4b50-92d5-5f7dabf6a55b/setup-container/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.040927 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_00618566-3b37-4b50-92d5-5f7dabf6a55b/rabbitmq/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.081762 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-qtnfd_31438de3-4165-4a60-8a13-bb24b1743cd6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.245867 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fqbcg_c6da5217-1cd1-4f4e-8636-bb4a3604e939/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.297863 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kshfl_234d0af6-f029-4b29-8003-b954e9e4e07a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.350106 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cnpqb_f98da165-e7ef-4475-84db-01079d69bb82/ssh-known-hosts-edpm-deployment/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.597621 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_c6b4d79c-6eea-4a14-bea9-5074054b09ca/tempest-tests-tempest-tests-runner/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.758501 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_928c3f00-4054-4f1d-8693-102b6032a11a/test-operator-logs-container/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.835125 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_c50bbf30-2132-45d9-93fb-4d266d85bf00/test-operator-logs-container/0.log" Nov 24 21:15:47 crc kubenswrapper[5035]: I1124 21:15:47.971933 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b504ae82-14cc-45d4-8eac-9c49b8ab7172/test-operator-logs-container/0.log" Nov 24 21:15:48 crc kubenswrapper[5035]: I1124 21:15:48.113020 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_b2081aca-59eb-43e9-b9fe-753837e7e553/test-operator-logs-container/0.log" Nov 24 21:15:48 crc kubenswrapper[5035]: I1124 21:15:48.241398 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_3b0c51fa-ab78-4018-a361-f02d53b92345/tobiko-tests-tobiko/0.log" Nov 24 21:15:48 crc kubenswrapper[5035]: I1124 21:15:48.445088 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-g2pkw_20cbe453-ecde-4214-835e-a2a6a1fb2cc3/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 21:15:48 crc kubenswrapper[5035]: I1124 21:15:48.556131 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_7af30223-20cb-4470-b630-170318d5de9e/tobiko-tests-tobiko/0.log" Nov 24 21:15:48 crc kubenswrapper[5035]: I1124 21:15:48.687752 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_03d6dde6-e977-4ec1-a23d-1642e6c23029/tempest-tests-tempest-tests-runner/0.log" Nov 24 21:15:49 crc kubenswrapper[5035]: I1124 21:15:49.199557 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:15:49 crc kubenswrapper[5035]: E1124 21:15:49.199808 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:16:02 crc kubenswrapper[5035]: I1124 21:16:02.211184 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:16:02 crc kubenswrapper[5035]: E1124 21:16:02.212103 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:16:10 crc kubenswrapper[5035]: I1124 21:16:10.865675 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-6rjkn_f779482e-56ae-49de-8d6d-e06bf57dc3a7/kube-rbac-proxy/0.log" Nov 24 21:16:10 crc kubenswrapper[5035]: I1124 21:16:10.992663 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-6rjkn_f779482e-56ae-49de-8d6d-e06bf57dc3a7/manager/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.222711 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp_089631e1-99a1-47f0-ac8f-d42bfb4f2b9f/util/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.524376 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp_089631e1-99a1-47f0-ac8f-d42bfb4f2b9f/util/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.592489 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp_089631e1-99a1-47f0-ac8f-d42bfb4f2b9f/pull/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.592564 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp_089631e1-99a1-47f0-ac8f-d42bfb4f2b9f/pull/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.769341 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp_089631e1-99a1-47f0-ac8f-d42bfb4f2b9f/util/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.789985 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp_089631e1-99a1-47f0-ac8f-d42bfb4f2b9f/pull/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.796409 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15flzggp_089631e1-99a1-47f0-ac8f-d42bfb4f2b9f/extract/0.log" Nov 24 21:16:11 crc kubenswrapper[5035]: I1124 21:16:11.935953 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-wp2d7_95f16522-f695-414f-8ec6-c2cb86913080/kube-rbac-proxy/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.003256 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-wp2d7_95f16522-f695-414f-8ec6-c2cb86913080/manager/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.021361 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-p8vr6_f4b3bf02-333f-46c2-90a0-fe92b166328c/kube-rbac-proxy/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.141679 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-p8vr6_f4b3bf02-333f-46c2-90a0-fe92b166328c/manager/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.175661 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-ws7tg_d99686fc-e38b-49de-89a3-7de31352efc3/kube-rbac-proxy/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.277704 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-ws7tg_d99686fc-e38b-49de-89a3-7de31352efc3/manager/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.355175 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-ckj7f_9ae5fef4-4b29-4ea2-8404-b145d960ef4a/kube-rbac-proxy/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.372125 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-ckj7f_9ae5fef4-4b29-4ea2-8404-b145d960ef4a/manager/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.545567 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-mq8nd_0f776395-f2dd-41cc-af5f-e38dcd9da7b3/manager/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.547009 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-mq8nd_0f776395-f2dd-41cc-af5f-e38dcd9da7b3/kube-rbac-proxy/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.660237 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-bvfmn_c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f/kube-rbac-proxy/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.744566 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-rd2qm_6084f8a1-4cde-4714-af8d-76f274dd4b68/kube-rbac-proxy/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.820822 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-bvfmn_c1a5df31-aeb5-4783-a5b7-b4d3ec37cd5f/manager/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.846162 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-rd2qm_6084f8a1-4cde-4714-af8d-76f274dd4b68/manager/0.log" Nov 24 21:16:12 crc kubenswrapper[5035]: I1124 21:16:12.954609 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-tfjff_af7760bb-5727-4a40-b9a8-9c7b48a9d066/kube-rbac-proxy/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.085848 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-tfjff_af7760bb-5727-4a40-b9a8-9c7b48a9d066/manager/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.143185 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-ln7hm_f835e1e6-2181-4243-9572-0566b72f2f11/kube-rbac-proxy/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.211209 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-ln7hm_f835e1e6-2181-4243-9572-0566b72f2f11/manager/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.303880 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-csfwz_792f458f-945a-4846-ac66-c3e2801c2be6/kube-rbac-proxy/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.339591 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-csfwz_792f458f-945a-4846-ac66-c3e2801c2be6/manager/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.436892 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-5wtlz_6db4802c-aa26-4fc1-baa7-37ac2845a460/kube-rbac-proxy/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.569995 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-5wtlz_6db4802c-aa26-4fc1-baa7-37ac2845a460/manager/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.646694 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-kkcw8_bf589985-cf9b-4133-8d16-794c8fbfa0f3/kube-rbac-proxy/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.747798 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-kkcw8_bf589985-cf9b-4133-8d16-794c8fbfa0f3/manager/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.791652 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-rqf69_00919c94-bf31-4ea8-a322-5b81bdb050c9/kube-rbac-proxy/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.824530 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-rqf69_00919c94-bf31-4ea8-a322-5b81bdb050c9/manager/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.937433 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-gqwwm_b10f4358-81d9-4d88-a0c5-3aa698f00cdf/kube-rbac-proxy/0.log" Nov 24 21:16:13 crc kubenswrapper[5035]: I1124 21:16:13.982035 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-gqwwm_b10f4358-81d9-4d88-a0c5-3aa698f00cdf/manager/0.log" Nov 24 21:16:14 crc kubenswrapper[5035]: I1124 21:16:14.213598 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-556f94c7bb-bdq4q_b9888ed0-83e5-418d-9187-564b3e4832eb/operator/0.log" Nov 24 21:16:14 crc kubenswrapper[5035]: I1124 21:16:14.541931 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-prr8k_ba2f4f9d-09b7-4192-9061-39d5204dd054/kube-rbac-proxy/0.log" Nov 24 21:16:14 crc kubenswrapper[5035]: I1124 21:16:14.548522 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-lkrfx_25e32e21-6ff4-4b8f-aa65-23c9cff7b8c9/registry-server/0.log" Nov 24 21:16:14 crc kubenswrapper[5035]: I1124 21:16:14.797903 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-prr8k_ba2f4f9d-09b7-4192-9061-39d5204dd054/manager/0.log" Nov 24 21:16:14 crc kubenswrapper[5035]: I1124 21:16:14.801588 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-87rk8_8ba349a2-1cc6-43ef-9045-4d02feffffd9/kube-rbac-proxy/0.log" Nov 24 21:16:14 crc kubenswrapper[5035]: I1124 21:16:14.854038 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-87rk8_8ba349a2-1cc6-43ef-9045-4d02feffffd9/manager/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.050284 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-n2s8q_98c2e05e-f4af-4979-960b-4104a843321a/operator/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.126479 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-tpw7d_be99a698-bf30-4dd1-b534-e785a37143bc/kube-rbac-proxy/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.286631 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-tpw7d_be99a698-bf30-4dd1-b534-e785a37143bc/manager/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.393059 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-grqcw_eab64b25-1263-4d64-ba86-1f7227feae79/kube-rbac-proxy/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.420107 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cd5954d9-hvjd4_4eb72ce7-5f41-4360-91be-e7bce1db7b69/manager/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.495971 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-grqcw_eab64b25-1263-4d64-ba86-1f7227feae79/manager/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.536143 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69bf7c8699-f9hjp_b675c0a3-b86b-4993-902e-7a6f041a907a/kube-rbac-proxy/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.578479 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69bf7c8699-f9hjp_b675c0a3-b86b-4993-902e-7a6f041a907a/manager/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.694838 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-rjs7d_3d6e2a89-1911-4c26-8a9c-03c9955af913/kube-rbac-proxy/0.log" Nov 24 21:16:15 crc kubenswrapper[5035]: I1124 21:16:15.723403 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-rjs7d_3d6e2a89-1911-4c26-8a9c-03c9955af913/manager/0.log" Nov 24 21:16:17 crc kubenswrapper[5035]: I1124 21:16:17.200388 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:16:17 crc kubenswrapper[5035]: E1124 21:16:17.200908 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:16:28 crc kubenswrapper[5035]: I1124 21:16:28.200104 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:16:28 crc kubenswrapper[5035]: E1124 21:16:28.200939 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:16:32 crc kubenswrapper[5035]: I1124 21:16:32.255220 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6r9rt_8105a388-83c1-4372-bd51-fcbc2a2e9198/control-plane-machine-set-operator/0.log" Nov 24 21:16:32 crc kubenswrapper[5035]: I1124 21:16:32.435630 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lwfhz_236d3c3c-fe73-409f-b39c-7bb3ba97c902/kube-rbac-proxy/0.log" Nov 24 21:16:32 crc kubenswrapper[5035]: I1124 21:16:32.476074 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-lwfhz_236d3c3c-fe73-409f-b39c-7bb3ba97c902/machine-api-operator/0.log" Nov 24 21:16:43 crc kubenswrapper[5035]: I1124 21:16:43.200738 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:16:43 crc kubenswrapper[5035]: E1124 21:16:43.201539 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:16:44 crc kubenswrapper[5035]: I1124 21:16:44.201983 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nxkrb_d1e84c7b-08b8-411b-af66-5b1dd7f7f520/cert-manager-controller/0.log" Nov 24 21:16:44 crc kubenswrapper[5035]: I1124 21:16:44.372870 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qbfvl_12d8a7dc-4150-452b-8c34-121cb6e65810/cert-manager-cainjector/0.log" Nov 24 21:16:44 crc kubenswrapper[5035]: I1124 21:16:44.388200 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-kktnk_2dc727b8-eefd-4ce5-bae3-ab99280b162a/cert-manager-webhook/0.log" Nov 24 21:16:55 crc kubenswrapper[5035]: I1124 21:16:55.766803 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-gz87f_b9793c81-a455-4d46-bc0a-768b57675a89/nmstate-console-plugin/0.log" Nov 24 21:16:55 crc kubenswrapper[5035]: I1124 21:16:55.964641 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vg29s_c3173bd9-7b88-45d8-bb49-197c85649bc1/nmstate-handler/0.log" Nov 24 21:16:56 crc kubenswrapper[5035]: I1124 21:16:56.015144 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-rg4vj_02230763-8196-4df6-8d05-0e6940217381/kube-rbac-proxy/0.log" Nov 24 21:16:56 crc kubenswrapper[5035]: I1124 21:16:56.033641 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-rg4vj_02230763-8196-4df6-8d05-0e6940217381/nmstate-metrics/0.log" Nov 24 21:16:56 crc kubenswrapper[5035]: I1124 21:16:56.201014 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-cdj9v_b0ce9ef4-de1c-41fa-b76f-2a31fc591d54/nmstate-operator/0.log" Nov 24 21:16:56 crc kubenswrapper[5035]: I1124 21:16:56.205893 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:16:56 crc kubenswrapper[5035]: E1124 21:16:56.206442 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:16:56 crc kubenswrapper[5035]: I1124 21:16:56.252941 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-fxwc7_74f8a61e-f6be-4f79-9eb9-7a196d884e29/nmstate-webhook/0.log" Nov 24 21:17:00 crc kubenswrapper[5035]: E1124 21:17:00.200431 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:17:07 crc kubenswrapper[5035]: I1124 21:17:07.199768 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:17:07 crc kubenswrapper[5035]: E1124 21:17:07.200572 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:17:11 crc kubenswrapper[5035]: I1124 21:17:11.964066 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-xpd9f_785073c0-865d-4f23-814a-a393cd4caced/kube-rbac-proxy/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.035150 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-xpd9f_785073c0-865d-4f23-814a-a393cd4caced/controller/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.099572 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-frr-files/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.283144 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-reloader/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.301657 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-frr-files/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.302342 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-reloader/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.359426 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-metrics/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.511411 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-reloader/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.513525 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-frr-files/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.523306 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-metrics/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.542347 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-metrics/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.704095 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-frr-files/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.704216 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-reloader/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.735934 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/cp-metrics/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.741535 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/controller/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.926271 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/frr-metrics/0.log" Nov 24 21:17:12 crc kubenswrapper[5035]: I1124 21:17:12.982329 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/kube-rbac-proxy/0.log" Nov 24 21:17:13 crc kubenswrapper[5035]: I1124 21:17:13.034784 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/kube-rbac-proxy-frr/0.log" Nov 24 21:17:13 crc kubenswrapper[5035]: I1124 21:17:13.189519 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/reloader/0.log" Nov 24 21:17:13 crc kubenswrapper[5035]: I1124 21:17:13.300256 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-9drwf_4041fa2f-450a-4f60-baa7-15a8478934f4/frr-k8s-webhook-server/0.log" Nov 24 21:17:13 crc kubenswrapper[5035]: I1124 21:17:13.402236 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6648cb8994-hqcpk_b782ab87-baed-4a09-b133-affcf2efd744/manager/0.log" Nov 24 21:17:13 crc kubenswrapper[5035]: I1124 21:17:13.620984 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6fbc66c766-4477t_60ba6695-0ae0-4dce-b6de-ec96a4723d07/webhook-server/0.log" Nov 24 21:17:13 crc kubenswrapper[5035]: I1124 21:17:13.809141 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xxqk2_4ae03b6d-7158-40a9-9b24-4a247860e4e1/kube-rbac-proxy/0.log" Nov 24 21:17:14 crc kubenswrapper[5035]: I1124 21:17:14.398445 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xxqk2_4ae03b6d-7158-40a9-9b24-4a247860e4e1/speaker/0.log" Nov 24 21:17:14 crc kubenswrapper[5035]: I1124 21:17:14.885463 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jx6sg_9003e7e1-21c5-41e2-8eff-a55dbeabaaa0/frr/0.log" Nov 24 21:17:20 crc kubenswrapper[5035]: I1124 21:17:20.201903 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:17:20 crc kubenswrapper[5035]: E1124 21:17:20.202876 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.402394 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr_efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc/util/0.log" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.521758 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr_efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc/util/0.log" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.561030 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr_efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc/pull/0.log" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.576063 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr_efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc/pull/0.log" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.734945 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr_efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc/extract/0.log" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.735521 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr_efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc/pull/0.log" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.743900 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772eggmpr_efd76c4e-b7f8-4393-b3ef-1be1ffc8eedc/util/0.log" Nov 24 21:17:26 crc kubenswrapper[5035]: I1124 21:17:26.902454 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rkdl9_4faf7893-a5a7-4f59-b03c-b2c95642f103/extract-utilities/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.075503 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rkdl9_4faf7893-a5a7-4f59-b03c-b2c95642f103/extract-utilities/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.077046 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rkdl9_4faf7893-a5a7-4f59-b03c-b2c95642f103/extract-content/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.160960 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rkdl9_4faf7893-a5a7-4f59-b03c-b2c95642f103/extract-content/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.238942 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rkdl9_4faf7893-a5a7-4f59-b03c-b2c95642f103/extract-content/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.247847 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rkdl9_4faf7893-a5a7-4f59-b03c-b2c95642f103/extract-utilities/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.445224 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr6hr_14a39f18-63d0-4f2d-a1d8-e8117872b9f6/extract-utilities/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.691919 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rkdl9_4faf7893-a5a7-4f59-b03c-b2c95642f103/registry-server/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.734393 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr6hr_14a39f18-63d0-4f2d-a1d8-e8117872b9f6/extract-content/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.737171 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr6hr_14a39f18-63d0-4f2d-a1d8-e8117872b9f6/extract-utilities/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.758369 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr6hr_14a39f18-63d0-4f2d-a1d8-e8117872b9f6/extract-content/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.938254 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr6hr_14a39f18-63d0-4f2d-a1d8-e8117872b9f6/extract-utilities/0.log" Nov 24 21:17:27 crc kubenswrapper[5035]: I1124 21:17:27.959523 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr6hr_14a39f18-63d0-4f2d-a1d8-e8117872b9f6/extract-content/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.169609 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4_a606927e-ed47-44e2-a593-8dc09c393d71/util/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.421130 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4_a606927e-ed47-44e2-a593-8dc09c393d71/util/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.436026 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4_a606927e-ed47-44e2-a593-8dc09c393d71/pull/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.501495 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4_a606927e-ed47-44e2-a593-8dc09c393d71/pull/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.642385 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4_a606927e-ed47-44e2-a593-8dc09c393d71/util/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.643300 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4_a606927e-ed47-44e2-a593-8dc09c393d71/pull/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.719318 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6xq8g4_a606927e-ed47-44e2-a593-8dc09c393d71/extract/0.log" Nov 24 21:17:28 crc kubenswrapper[5035]: I1124 21:17:28.878767 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-49zbg_70cbf730-06f3-4b21-aa8c-64a80ae4929b/marketplace-operator/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.071029 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x2vrj_e5328a7d-e2f7-43ae-89b9-c264960c8912/extract-utilities/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.074792 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr6hr_14a39f18-63d0-4f2d-a1d8-e8117872b9f6/registry-server/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.249093 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x2vrj_e5328a7d-e2f7-43ae-89b9-c264960c8912/extract-content/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.276537 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x2vrj_e5328a7d-e2f7-43ae-89b9-c264960c8912/extract-utilities/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.304035 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x2vrj_e5328a7d-e2f7-43ae-89b9-c264960c8912/extract-content/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.454340 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x2vrj_e5328a7d-e2f7-43ae-89b9-c264960c8912/extract-content/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.491120 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x2vrj_e5328a7d-e2f7-43ae-89b9-c264960c8912/extract-utilities/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.718689 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x2vrj_e5328a7d-e2f7-43ae-89b9-c264960c8912/registry-server/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.722272 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2z6cz_52a63a97-8196-4344-bec6-df0c09b57642/extract-utilities/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.863550 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2z6cz_52a63a97-8196-4344-bec6-df0c09b57642/extract-utilities/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.903578 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2z6cz_52a63a97-8196-4344-bec6-df0c09b57642/extract-content/0.log" Nov 24 21:17:29 crc kubenswrapper[5035]: I1124 21:17:29.945448 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2z6cz_52a63a97-8196-4344-bec6-df0c09b57642/extract-content/0.log" Nov 24 21:17:30 crc kubenswrapper[5035]: I1124 21:17:30.137106 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2z6cz_52a63a97-8196-4344-bec6-df0c09b57642/extract-content/0.log" Nov 24 21:17:30 crc kubenswrapper[5035]: I1124 21:17:30.178363 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2z6cz_52a63a97-8196-4344-bec6-df0c09b57642/extract-utilities/0.log" Nov 24 21:17:30 crc kubenswrapper[5035]: I1124 21:17:30.830333 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2z6cz_52a63a97-8196-4344-bec6-df0c09b57642/registry-server/0.log" Nov 24 21:17:33 crc kubenswrapper[5035]: I1124 21:17:33.199952 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:17:33 crc kubenswrapper[5035]: E1124 21:17:33.200466 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:17:47 crc kubenswrapper[5035]: I1124 21:17:47.200245 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:17:47 crc kubenswrapper[5035]: I1124 21:17:47.508678 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"6c94754067d098056f23455771a60b7fa03f479170f68ad90fd927555664f811"} Nov 24 21:18:05 crc kubenswrapper[5035]: E1124 21:18:05.404236 5035 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.47:41776->38.102.83.47:43781: write tcp 38.102.83.47:41776->38.102.83.47:43781: write: broken pipe Nov 24 21:18:05 crc kubenswrapper[5035]: E1124 21:18:05.407863 5035 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.47:41730->38.102.83.47:43781: write tcp 38.102.83.47:41730->38.102.83.47:43781: write: broken pipe Nov 24 21:18:09 crc kubenswrapper[5035]: E1124 21:18:09.200768 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:19:37 crc kubenswrapper[5035]: I1124 21:19:37.672086 5035 generic.go:334] "Generic (PLEG): container finished" podID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerID="617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac" exitCode=0 Nov 24 21:19:37 crc kubenswrapper[5035]: I1124 21:19:37.672180 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pw2tc/must-gather-l46g5" event={"ID":"6b1c38fb-066c-4261-8e53-a38f6e3ac622","Type":"ContainerDied","Data":"617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac"} Nov 24 21:19:37 crc kubenswrapper[5035]: I1124 21:19:37.673357 5035 scope.go:117] "RemoveContainer" containerID="617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac" Nov 24 21:19:38 crc kubenswrapper[5035]: I1124 21:19:38.140219 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pw2tc_must-gather-l46g5_6b1c38fb-066c-4261-8e53-a38f6e3ac622/gather/0.log" Nov 24 21:19:39 crc kubenswrapper[5035]: E1124 21:19:39.201063 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.211529 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pw2tc/must-gather-l46g5"] Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.212219 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pw2tc/must-gather-l46g5" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerName="copy" containerID="cri-o://9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982" gracePeriod=2 Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.240818 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pw2tc/must-gather-l46g5"] Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.653994 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pw2tc_must-gather-l46g5_6b1c38fb-066c-4261-8e53-a38f6e3ac622/copy/0.log" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.654700 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.748447 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b1c38fb-066c-4261-8e53-a38f6e3ac622-must-gather-output\") pod \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.748560 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pfkk\" (UniqueName: \"kubernetes.io/projected/6b1c38fb-066c-4261-8e53-a38f6e3ac622-kube-api-access-6pfkk\") pod \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\" (UID: \"6b1c38fb-066c-4261-8e53-a38f6e3ac622\") " Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.754755 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b1c38fb-066c-4261-8e53-a38f6e3ac622-kube-api-access-6pfkk" (OuterVolumeSpecName: "kube-api-access-6pfkk") pod "6b1c38fb-066c-4261-8e53-a38f6e3ac622" (UID: "6b1c38fb-066c-4261-8e53-a38f6e3ac622"). InnerVolumeSpecName "kube-api-access-6pfkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.763313 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pw2tc_must-gather-l46g5_6b1c38fb-066c-4261-8e53-a38f6e3ac622/copy/0.log" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.763734 5035 generic.go:334] "Generic (PLEG): container finished" podID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerID="9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982" exitCode=143 Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.763779 5035 scope.go:117] "RemoveContainer" containerID="9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.763892 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pw2tc/must-gather-l46g5" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.812158 5035 scope.go:117] "RemoveContainer" containerID="617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.857811 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pfkk\" (UniqueName: \"kubernetes.io/projected/6b1c38fb-066c-4261-8e53-a38f6e3ac622-kube-api-access-6pfkk\") on node \"crc\" DevicePath \"\"" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.892108 5035 scope.go:117] "RemoveContainer" containerID="9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982" Nov 24 21:19:46 crc kubenswrapper[5035]: E1124 21:19:46.899449 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982\": container with ID starting with 9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982 not found: ID does not exist" containerID="9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.899499 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982"} err="failed to get container status \"9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982\": rpc error: code = NotFound desc = could not find container \"9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982\": container with ID starting with 9c597430d4bf1b37cc0468008134cb4f85479f55a97ae43750924f885f2c7982 not found: ID does not exist" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.899525 5035 scope.go:117] "RemoveContainer" containerID="617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac" Nov 24 21:19:46 crc kubenswrapper[5035]: E1124 21:19:46.903439 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac\": container with ID starting with 617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac not found: ID does not exist" containerID="617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.903486 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac"} err="failed to get container status \"617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac\": rpc error: code = NotFound desc = could not find container \"617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac\": container with ID starting with 617eaaa9a26f614c95632b2233cff3447cba2b6de96b4d1aefb9eb8b0d6f95ac not found: ID does not exist" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.909634 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b1c38fb-066c-4261-8e53-a38f6e3ac622-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6b1c38fb-066c-4261-8e53-a38f6e3ac622" (UID: "6b1c38fb-066c-4261-8e53-a38f6e3ac622"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:19:46 crc kubenswrapper[5035]: I1124 21:19:46.960072 5035 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b1c38fb-066c-4261-8e53-a38f6e3ac622-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 21:19:48 crc kubenswrapper[5035]: I1124 21:19:48.209789 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" path="/var/lib/kubelet/pods/6b1c38fb-066c-4261-8e53-a38f6e3ac622/volumes" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.548425 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vb86q"] Nov 24 21:20:06 crc kubenswrapper[5035]: E1124 21:20:06.549468 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerName="copy" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.549484 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerName="copy" Nov 24 21:20:06 crc kubenswrapper[5035]: E1124 21:20:06.549523 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e750574-8891-4edc-b063-b3952b3f42d7" containerName="container-00" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.549531 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e750574-8891-4edc-b063-b3952b3f42d7" containerName="container-00" Nov 24 21:20:06 crc kubenswrapper[5035]: E1124 21:20:06.549551 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerName="gather" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.549558 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerName="gather" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.549780 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e750574-8891-4edc-b063-b3952b3f42d7" containerName="container-00" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.549799 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerName="gather" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.549814 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1c38fb-066c-4261-8e53-a38f6e3ac622" containerName="copy" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.551438 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.557471 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vb86q"] Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.577314 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2485p\" (UniqueName: \"kubernetes.io/projected/fa98fb49-76f6-482a-90fa-3dab07016473-kube-api-access-2485p\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.577382 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-catalog-content\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.577628 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-utilities\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.679102 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2485p\" (UniqueName: \"kubernetes.io/projected/fa98fb49-76f6-482a-90fa-3dab07016473-kube-api-access-2485p\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.679176 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-catalog-content\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.679222 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-utilities\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.679779 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-utilities\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.679892 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-catalog-content\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.707107 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2485p\" (UniqueName: \"kubernetes.io/projected/fa98fb49-76f6-482a-90fa-3dab07016473-kube-api-access-2485p\") pod \"community-operators-vb86q\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:06 crc kubenswrapper[5035]: I1124 21:20:06.873518 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:07 crc kubenswrapper[5035]: I1124 21:20:07.441979 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vb86q"] Nov 24 21:20:07 crc kubenswrapper[5035]: I1124 21:20:07.992429 5035 generic.go:334] "Generic (PLEG): container finished" podID="fa98fb49-76f6-482a-90fa-3dab07016473" containerID="ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94" exitCode=0 Nov 24 21:20:07 crc kubenswrapper[5035]: I1124 21:20:07.992493 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vb86q" event={"ID":"fa98fb49-76f6-482a-90fa-3dab07016473","Type":"ContainerDied","Data":"ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94"} Nov 24 21:20:07 crc kubenswrapper[5035]: I1124 21:20:07.993966 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vb86q" event={"ID":"fa98fb49-76f6-482a-90fa-3dab07016473","Type":"ContainerStarted","Data":"e83f533532edaf18e1aa9eec0ab2831470860c61af363ca819b50fc6faf60dbf"} Nov 24 21:20:07 crc kubenswrapper[5035]: I1124 21:20:07.995083 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 21:20:10 crc kubenswrapper[5035]: I1124 21:20:10.018520 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vb86q" event={"ID":"fa98fb49-76f6-482a-90fa-3dab07016473","Type":"ContainerStarted","Data":"8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747"} Nov 24 21:20:11 crc kubenswrapper[5035]: I1124 21:20:11.034221 5035 generic.go:334] "Generic (PLEG): container finished" podID="fa98fb49-76f6-482a-90fa-3dab07016473" containerID="8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747" exitCode=0 Nov 24 21:20:11 crc kubenswrapper[5035]: I1124 21:20:11.034515 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vb86q" event={"ID":"fa98fb49-76f6-482a-90fa-3dab07016473","Type":"ContainerDied","Data":"8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747"} Nov 24 21:20:12 crc kubenswrapper[5035]: I1124 21:20:12.048876 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vb86q" event={"ID":"fa98fb49-76f6-482a-90fa-3dab07016473","Type":"ContainerStarted","Data":"01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497"} Nov 24 21:20:12 crc kubenswrapper[5035]: I1124 21:20:12.080100 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vb86q" podStartSLOduration=2.590333288 podStartE2EDuration="6.080070766s" podCreationTimestamp="2025-11-24 21:20:06 +0000 UTC" firstStartedPulling="2025-11-24 21:20:07.994800546 +0000 UTC m=+7546.517306803" lastFinishedPulling="2025-11-24 21:20:11.484538014 +0000 UTC m=+7550.007044281" observedRunningTime="2025-11-24 21:20:12.075200666 +0000 UTC m=+7550.597706933" watchObservedRunningTime="2025-11-24 21:20:12.080070766 +0000 UTC m=+7550.602577023" Nov 24 21:20:15 crc kubenswrapper[5035]: I1124 21:20:15.234609 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:20:15 crc kubenswrapper[5035]: I1124 21:20:15.234983 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:20:16 crc kubenswrapper[5035]: I1124 21:20:16.874620 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:16 crc kubenswrapper[5035]: I1124 21:20:16.874689 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:16 crc kubenswrapper[5035]: I1124 21:20:16.944421 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:17 crc kubenswrapper[5035]: I1124 21:20:17.158586 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:17 crc kubenswrapper[5035]: I1124 21:20:17.217708 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vb86q"] Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.129786 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vb86q" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="registry-server" containerID="cri-o://01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497" gracePeriod=2 Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.616925 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.667263 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-utilities\") pod \"fa98fb49-76f6-482a-90fa-3dab07016473\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.667480 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-catalog-content\") pod \"fa98fb49-76f6-482a-90fa-3dab07016473\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.667527 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2485p\" (UniqueName: \"kubernetes.io/projected/fa98fb49-76f6-482a-90fa-3dab07016473-kube-api-access-2485p\") pod \"fa98fb49-76f6-482a-90fa-3dab07016473\" (UID: \"fa98fb49-76f6-482a-90fa-3dab07016473\") " Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.669541 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-utilities" (OuterVolumeSpecName: "utilities") pod "fa98fb49-76f6-482a-90fa-3dab07016473" (UID: "fa98fb49-76f6-482a-90fa-3dab07016473"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.696506 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa98fb49-76f6-482a-90fa-3dab07016473-kube-api-access-2485p" (OuterVolumeSpecName: "kube-api-access-2485p") pod "fa98fb49-76f6-482a-90fa-3dab07016473" (UID: "fa98fb49-76f6-482a-90fa-3dab07016473"). InnerVolumeSpecName "kube-api-access-2485p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.772790 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.772837 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2485p\" (UniqueName: \"kubernetes.io/projected/fa98fb49-76f6-482a-90fa-3dab07016473-kube-api-access-2485p\") on node \"crc\" DevicePath \"\"" Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.967817 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa98fb49-76f6-482a-90fa-3dab07016473" (UID: "fa98fb49-76f6-482a-90fa-3dab07016473"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 21:20:19 crc kubenswrapper[5035]: I1124 21:20:19.977190 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa98fb49-76f6-482a-90fa-3dab07016473-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.142631 5035 generic.go:334] "Generic (PLEG): container finished" podID="fa98fb49-76f6-482a-90fa-3dab07016473" containerID="01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497" exitCode=0 Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.142701 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vb86q" event={"ID":"fa98fb49-76f6-482a-90fa-3dab07016473","Type":"ContainerDied","Data":"01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497"} Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.142721 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vb86q" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.142793 5035 scope.go:117] "RemoveContainer" containerID="01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.142776 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vb86q" event={"ID":"fa98fb49-76f6-482a-90fa-3dab07016473","Type":"ContainerDied","Data":"e83f533532edaf18e1aa9eec0ab2831470860c61af363ca819b50fc6faf60dbf"} Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.175448 5035 scope.go:117] "RemoveContainer" containerID="8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.212024 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vb86q"] Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.223893 5035 scope.go:117] "RemoveContainer" containerID="ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.224136 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vb86q"] Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.269074 5035 scope.go:117] "RemoveContainer" containerID="01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497" Nov 24 21:20:20 crc kubenswrapper[5035]: E1124 21:20:20.269535 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497\": container with ID starting with 01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497 not found: ID does not exist" containerID="01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.269582 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497"} err="failed to get container status \"01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497\": rpc error: code = NotFound desc = could not find container \"01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497\": container with ID starting with 01592760405da33ce97ac9aea7e7eb4639068aa845a1ed72e4aaf791fbc0f497 not found: ID does not exist" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.269607 5035 scope.go:117] "RemoveContainer" containerID="8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747" Nov 24 21:20:20 crc kubenswrapper[5035]: E1124 21:20:20.270057 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747\": container with ID starting with 8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747 not found: ID does not exist" containerID="8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.270084 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747"} err="failed to get container status \"8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747\": rpc error: code = NotFound desc = could not find container \"8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747\": container with ID starting with 8b2187198e93b871abed05c5f0ec9127a3fc85cc85cd7b13b6980c951e082747 not found: ID does not exist" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.270098 5035 scope.go:117] "RemoveContainer" containerID="ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94" Nov 24 21:20:20 crc kubenswrapper[5035]: E1124 21:20:20.270455 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94\": container with ID starting with ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94 not found: ID does not exist" containerID="ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94" Nov 24 21:20:20 crc kubenswrapper[5035]: I1124 21:20:20.270493 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94"} err="failed to get container status \"ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94\": rpc error: code = NotFound desc = could not find container \"ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94\": container with ID starting with ff077177cd51ce48d70da34a1e6e0a09eed060e5c1ee12b1f46b83ce6befbc94 not found: ID does not exist" Nov 24 21:20:22 crc kubenswrapper[5035]: I1124 21:20:22.215526 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" path="/var/lib/kubelet/pods/fa98fb49-76f6-482a-90fa-3dab07016473/volumes" Nov 24 21:20:41 crc kubenswrapper[5035]: E1124 21:20:41.200527 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:20:45 crc kubenswrapper[5035]: I1124 21:20:45.234630 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:20:45 crc kubenswrapper[5035]: I1124 21:20:45.235229 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.234683 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.235352 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.235400 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.236178 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c94754067d098056f23455771a60b7fa03f479170f68ad90fd927555664f811"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.236250 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://6c94754067d098056f23455771a60b7fa03f479170f68ad90fd927555664f811" gracePeriod=600 Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.678026 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="6c94754067d098056f23455771a60b7fa03f479170f68ad90fd927555664f811" exitCode=0 Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.678098 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"6c94754067d098056f23455771a60b7fa03f479170f68ad90fd927555664f811"} Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.678587 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerStarted","Data":"b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93"} Nov 24 21:21:15 crc kubenswrapper[5035]: I1124 21:21:15.678660 5035 scope.go:117] "RemoveContainer" containerID="32f6ba9e043922cd6417deffa9c0b6138efa55aaa88d6707a38fa0bb2f17759e" Nov 24 21:21:33 crc kubenswrapper[5035]: I1124 21:21:33.900978 5035 scope.go:117] "RemoveContainer" containerID="0af64bac3ebd609f82b9f250f5140b9d7409d8857f0a98c6b2be6457bd438f5c" Nov 24 21:21:50 crc kubenswrapper[5035]: E1124 21:21:50.200817 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:23:15 crc kubenswrapper[5035]: I1124 21:23:15.234541 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:23:15 crc kubenswrapper[5035]: I1124 21:23:15.235321 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:23:17 crc kubenswrapper[5035]: E1124 21:23:17.201969 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:23:45 crc kubenswrapper[5035]: I1124 21:23:45.234068 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:23:45 crc kubenswrapper[5035]: I1124 21:23:45.234772 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.233904 5035 patch_prober.go:28] interesting pod/machine-config-daemon-nvql4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.234377 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.234424 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.235194 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93"} pod="openshift-machine-config-operator/machine-config-daemon-nvql4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.235245 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerName="machine-config-daemon" containerID="cri-o://b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93" gracePeriod=600 Nov 24 21:24:15 crc kubenswrapper[5035]: E1124 21:24:15.357036 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.510196 5035 generic.go:334] "Generic (PLEG): container finished" podID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" containerID="b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93" exitCode=0 Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.510264 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" event={"ID":"11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361","Type":"ContainerDied","Data":"b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93"} Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.510365 5035 scope.go:117] "RemoveContainer" containerID="6c94754067d098056f23455771a60b7fa03f479170f68ad90fd927555664f811" Nov 24 21:24:15 crc kubenswrapper[5035]: I1124 21:24:15.511356 5035 scope.go:117] "RemoveContainer" containerID="b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93" Nov 24 21:24:15 crc kubenswrapper[5035]: E1124 21:24:15.511736 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:24:26 crc kubenswrapper[5035]: E1124 21:24:26.200477 5035 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Nov 24 21:24:27 crc kubenswrapper[5035]: I1124 21:24:27.201903 5035 scope.go:117] "RemoveContainer" containerID="b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93" Nov 24 21:24:27 crc kubenswrapper[5035]: E1124 21:24:27.202709 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:24:39 crc kubenswrapper[5035]: I1124 21:24:39.200371 5035 scope.go:117] "RemoveContainer" containerID="b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93" Nov 24 21:24:39 crc kubenswrapper[5035]: E1124 21:24:39.201573 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:24:51 crc kubenswrapper[5035]: I1124 21:24:51.200352 5035 scope.go:117] "RemoveContainer" containerID="b43e3192b21165979e8dbc7c8e176fbeb8cd622cd02eb97a66596a533a2b1e93" Nov 24 21:24:51 crc kubenswrapper[5035]: E1124 21:24:51.201235 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nvql4_openshift-machine-config-operator(11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361)\"" pod="openshift-machine-config-operator/machine-config-daemon-nvql4" podUID="11fbb8c9-66d2-4fdd-bb91-b4dfb6ea9361" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.466464 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mcj7w"] Nov 24 21:24:53 crc kubenswrapper[5035]: E1124 21:24:53.467441 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="extract-utilities" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.467463 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="extract-utilities" Nov 24 21:24:53 crc kubenswrapper[5035]: E1124 21:24:53.467490 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="registry-server" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.467502 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="registry-server" Nov 24 21:24:53 crc kubenswrapper[5035]: E1124 21:24:53.467553 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="extract-content" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.467564 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="extract-content" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.467894 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa98fb49-76f6-482a-90fa-3dab07016473" containerName="registry-server" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.470161 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.474541 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mcj7w"] Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.512618 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-catalog-content\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.513046 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2zsh\" (UniqueName: \"kubernetes.io/projected/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-kube-api-access-m2zsh\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.513356 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-utilities\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.615708 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-utilities\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.615853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-catalog-content\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.615880 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2zsh\" (UniqueName: \"kubernetes.io/projected/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-kube-api-access-m2zsh\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.616262 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-utilities\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.616325 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-catalog-content\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.640013 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2zsh\" (UniqueName: \"kubernetes.io/projected/65b5103d-95b4-4dd2-892b-a98ad6cfaeed-kube-api-access-m2zsh\") pod \"certified-operators-mcj7w\" (UID: \"65b5103d-95b4-4dd2-892b-a98ad6cfaeed\") " pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:53 crc kubenswrapper[5035]: I1124 21:24:53.800030 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mcj7w" Nov 24 21:24:54 crc kubenswrapper[5035]: I1124 21:24:54.325919 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mcj7w"] Nov 24 21:24:54 crc kubenswrapper[5035]: W1124 21:24:54.336502 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65b5103d_95b4_4dd2_892b_a98ad6cfaeed.slice/crio-64d6e3f1ae20bcc70af99050ac2c3439ba0344452fe2f4d9b06509376600f16b WatchSource:0}: Error finding container 64d6e3f1ae20bcc70af99050ac2c3439ba0344452fe2f4d9b06509376600f16b: Status 404 returned error can't find the container with id 64d6e3f1ae20bcc70af99050ac2c3439ba0344452fe2f4d9b06509376600f16b Nov 24 21:24:55 crc kubenswrapper[5035]: I1124 21:24:55.307235 5035 generic.go:334] "Generic (PLEG): container finished" podID="65b5103d-95b4-4dd2-892b-a98ad6cfaeed" containerID="a301dcb05e7c427c05095ff92cfcfdbf192f7e71aa5771a7510d7301e946b792" exitCode=0 Nov 24 21:24:55 crc kubenswrapper[5035]: I1124 21:24:55.307381 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mcj7w" event={"ID":"65b5103d-95b4-4dd2-892b-a98ad6cfaeed","Type":"ContainerDied","Data":"a301dcb05e7c427c05095ff92cfcfdbf192f7e71aa5771a7510d7301e946b792"} Nov 24 21:24:55 crc kubenswrapper[5035]: I1124 21:24:55.307585 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mcj7w" event={"ID":"65b5103d-95b4-4dd2-892b-a98ad6cfaeed","Type":"ContainerStarted","Data":"64d6e3f1ae20bcc70af99050ac2c3439ba0344452fe2f4d9b06509376600f16b"} Nov 24 21:24:56 crc kubenswrapper[5035]: I1124 21:24:56.320619 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mcj7w" event={"ID":"65b5103d-95b4-4dd2-892b-a98ad6cfaeed","Type":"ContainerStarted","Data":"208168d1a59119f1f131562552c7d99c05165db134c33e84607456991e3e77ef"} Nov 24 21:24:57 crc kubenswrapper[5035]: I1124 21:24:57.332882 5035 generic.go:334] "Generic (PLEG): container finished" podID="65b5103d-95b4-4dd2-892b-a98ad6cfaeed" containerID="208168d1a59119f1f131562552c7d99c05165db134c33e84607456991e3e77ef" exitCode=0 Nov 24 21:24:57 crc kubenswrapper[5035]: I1124 21:24:57.332972 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mcj7w" event={"ID":"65b5103d-95b4-4dd2-892b-a98ad6cfaeed","Type":"ContainerDied","Data":"208168d1a59119f1f131562552c7d99c05165db134c33e84607456991e3e77ef"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111146460024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111146461017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111126664016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111126665015461 5ustar corecore